Human Computer Interface HCI Seminar Report
computer science crazy|
Joined: Dec 2008
31-12-2009, 06:40 PM
Human Computer Interface HCI Seminar Report.doc (Size: 101 KB / Downloads: 474)
The Human-Computer Interface is a multidisciplinary field concerned with the design and evaluation of interacting computer systems from the user perspective. It deals with the way that computers and their users communicate. It pertains to the performance of tasks by humans and machines, the structure of human-machine communication, the engineering aspects of designing, building and specifying interfaces so that computers are easy to use and do what people want them to do. HCI requires the understanding of the capabilities and limitations of both sides of the communication. Information technologists should be aware of these issues in order to be able to produce software with interfaces that minimize user error and maximize user satisfaction.
The Human-Computer Interface (HCI) deals with the methods by which computers and their users communicate. It is the process of designing interface software so that computers are pleasant, easy to use and do what people want them to do. Dealing with HCI requires the study of not only the hardware of the computer, but that of the human side also. Therefore attention must be paid to human psychology and physiology. This is because to build a better two-way communication, one must know the capabilities and limitation of both sides. This seminar and presentation also deals with concepts and guidelines that should be followed in order to produce a good HCI. Specifically dealt with topics include Dialogue Design, Presentation Design, General Input and Output.
HUMAN PSYCHOLOGY & PHYSIOLOGY
This section mainly deals with the way humans communicate.
The human brain is where all the cognitive functions take place. It is ultimately where a human receives, interprets and stores information. Information can be processed by the sense organs and sent to the brain faster and more precise than the brain can handle. Many models have been developed in order to try and use a computer analogy to brain functions but with mixed success. They are however quite useful because they present to us a model with which we can illustrate capabilities and limitations.
These models suggest that there are two forms of human memory: short term and long term. Each sense appears to have its own short-term memory, which acts like a buffer or staging area for input from the particular sense organ to the brain. Any memory that is not reinforced and moved to long-term memory is forgotten. Short term memory has a capacity of about seven blocks of information but this too seems to be able to be increased with practice and added levels of abstraction and association.
In order for information to be remembered it must be moved into long-term memory. This can be a conscious act as in deliberately memorizing something through repetition or unconscious as when a particularly interesting piece of data is retrieved and requires more thought. No maximum size of long-term memory has yet been determined. This aspect of memory and the fact that the human brain can only process so much information is important to the layout of an HCI. People sometimes describe a particular screen as "too busy". What this means is that there is too much information on the screen at once. The brain is incapable of taking in so much information at once and ambiguity and confusion results. Precision should be a primary concern for the HCI designer.
Research has also indicated that a personâ„¢s ability to receive and interpret data can be influenced by outside factors like noise, lighting and other ergonomic factors. The operators stress and fatigue can impair their ability to function as well.
The primary method of receiving information from a computer is visually. The physiology of the eye will determine what limitations must be placed on an HCI. Current technology is able to present information at a faster rate then the eye can see. Limitations of HCI seem to be placed by the Human, not the Computer.
There are two different types of photoreceptors on the retina which are commonly referred to as "Rods" and "Cones". Rods are very sensitive to light whereas Cones are less sensitive. The angle of view for which there is a clear image is actually quite small, no more than 1 or 2 degrees. Clarity decreases rapidly in the optic field of view as one moves away from the center of the image. The amount of data that can be accurately seen within a single view is therefore also quite small and the eye must continually be moving in order to see a complete screen and then must mentally assimilate all the data into a complete mental page. This, in combination with the capacity of short-term memory sets a relatively low limit on the amount of data that can be contained on a single screen.
The eye's ability to distinguish colour, luminance, contrast and brightness also is a factor which must be considered in HCI. The colour of an object is determined by the frequency of the light that is reflected from it. Colour sensitivity is created by the existence of three different types of cones in the eye. One for blue, one for green and one for red. Each type of cone responds to a certain range of wavelengths. Colour blindness is the inability to distinguish certain colours, notably reds and greens. This fact is also important to remember when designing an HCI.
Luminance is a measure of the amount of light reflected from a surface. It is determined by the amount of light that is falling on an object and the reflectance of the surface of the object. Research has determined that there is a range of optimal luminance levels and that low illumination can be a hindrance to an otherwise good HCI.
Contrast is defined as the difference between the luminance of an object and its background divided by the luminance of the background. It is a measure of an eyes ability to distinguish foreground from background easily. It can have both a positive and a negative number. The higher the absolute value of the contrast the easier it is to distinguish objects. What this suggests is that in order for a user to be able to assimilate the information on a screen better, the background colour and the colour of the foreground text should be of opposing intensity. For example, black writing on a white background is ideal.
Brightness is usually thought of as a subjective property of light. It depends on many factors, the main one being comparative illumination. Brightness contrast can cause several common optical illusions as well. A common one is the "Hermann Grid" shown below as figure 1.0. Due to the comparative black and white sections, gray dots that aren't there are seen at the intersections.
Figure 1.0 - The Hermann Grid
The physical characteristics of the eyes and of vision must be taken into consideration when designing an HCI and knowledge of basic vision principles can greatly aid the HCI designer.
Although there is no question that visual is the primary means of relaying information from a computer to a human, practically every personal computer purchased today has a sound card and speakers. Audio is becoming a common way of relaying information to a user. Many help packages for software have an audio as well as visual component. Having a basic understanding of human hearing, capabilities and limitations can also aid the HCI designer in designing audio HCI components. Hearing involves the same problems inherent to seeing; reception of the signal, translating it into nerve impulses and attaching meaning to the messages. Sound waves travel through air. They have two major attributes which will be of concern to the HCI designer, frequency and amplitude. Frequency determines the pitch of the sound and amplitude determines the volume. Young children can hear in the range of about 20 Hz to over 15,000 Hz but this range decreases with age. Rarely are sounds of a single frequency however. Speech, as well as most everyday sounds, is a very complex mixture of frequencies.
What is more important than the anatomy of the ear and the physics of sound is the ear and brains ability to interpret sound. The auditory centre of the brain appears to be able to distinguish three different types of sound; background unimportant sounds (noise), background sounds that have significance (child's cry, dog's bark, etc.) and speech. The same sounds can therefore be "heard" differently depending on the context. Speech is a continuous thing. When analyzed, it doesn't appear as disjointed syllables or phonemes but as a continuous stream that must be interpreted at a rate of between 160 and 220 words per minute
The sense of touch is used as a primary input source and rarely as an output. Touch is mostly used for output when dealing with the disabled; sight and hearing impaired. Some pagers vibrate to attract the wearerâ„¢s attention etc. The use of a keyboard, mouse, joystick, light pen, touch pad etc. all use the sense of touch and the reason that there are so many different keyboards, mice etc. on the market is due to the different "touch" that each device has. The hand shaped mouse and the "split" keyboard are all examples of devices made to improve the touch of the device.
Vision is the primary sense used in HCI design and any designer should have a complete understanding of human vision capabilities and limitations. Hearing is the second most important sense in HCI design and anyone who undertakes a project and implimentation with an audio component should have an understanding of how the ear works and the method by which speech is interpreted in order to aid in design. Touch should be well understood for the design of hardware and special function devices.
INTERFACE ANALYSIS AND SPECIFICATION
Before one can design an appropriate HCI one has to know the reason for the interface. In other words, an HCI will be part of a larger project and implimentation developing a complete product for which the HCI is only a component. An HCI can only be designed if the purpose for the product itself is well known and understood. This is the task of Requirements Engineering in which the requirements for the system are elicited from the stakeholders, confirmed and documented in a manner such that a system can be developed from them.
There are several diverse techniques available for the Requirements Engineer to use in gathering requirements. One of the components of the results of these techniques will be an interface design. In order to create an interface design one must begin the task of "Functional Decomposition". Functional Decomposition is the task of examining a system or set of requirements and breaking it down into its individual components. Determine what base functionality you want the system to do. Document these singly and also in such a manner as to show the relationships between the different tasks. Also list each function using a spreadsheet or matrix. From this list of required functionalities, determine the ones that require an HCI. Before deciding the specifications of the HCI, the designer must look at the target market for the product. Four levels of users are generally defined:
1. Naive - These are the users who have never encountered or used a computer in their lives. With the fact that computers now permeate our lives to such an extent, this group is becoming quite small, but will always exist.
2. Novice - These are users who are slightly familiar with computers but would be quite unfamiliar with your system and how it would work. They are not computer phobic but just lack the exposure and experience.
3. Skilled - Skilled users have considerable computer experience and would be quite comfortable operating most computer systems. They would know where to go for help with a system functional problem but still would not have the expertise to understand the internal working of a computer or application.
4. Expert - The expert user is extremely comfortable with many aspects of computers and systems. They understand how they function and would attempt first to correct many problems themselves before obtaining professional help.
The majority of users today fall into categories 2 and 3. Caution must be taken when designing an HCI for a market containing a number of level 1 users.
As well as understanding the types of users that will eventually be using your product, it is equally important to understand a number of other user and system attributes. Consider the following user and application characteristics:
1. Frequency of use - How often will this system be used? Some systems, for example, are to be used only once and once used, discarded. A year 2000 fix program is a prime example. If the system is to be used only one a year, such as a system for closing year end financials, then the user cannot be expected to learn and retain system functionality from one use to the next. Many times it will be a new user each use.
2. Discretionary usage - An application may be used everyday by the same person so that workarounds, tricks, shortcuts etc. can be used quite frequently. What if the main user is away and someone who has never used this system suddenly has to step in?
3. User knowledge - A user may be a skilled computer user but totally unskilled in this particular application. You don't want to treat an expert user as a novice nor do you want to treat every user as an expert.
4. Mental abilities - It is stated that all systems should be "idiot proof". Too "idiot proof" can prevent a sophisticated user from making the best use out of the system.
5. Physical skills - Most applications assume a normal human that can see, hear, has two arms etc. A vast number of people don't have these capabilities and the area of developing interfaces for the handicapped in a large area of research.
Having determined what interfaces are required based on your requirements document and the type of user that you will be designing the interface for, one must look at the different components of an HCI. There are the two basic components, that of getting information into an application and retrieving it from the application. Most decisions made by an application require more than a single piece of data however. A dialogue must be created between the application and the user in order to retrieve all the required information, in the correct order and format, and this is necessary to have the application fulfill its function.
The intention of an HCI is to construct a mechanism so that a user can communicate with the desired application. A computer application is a tool and the HCI is the interface between the tool and the user. Most tools have interfaces of one form or another. For example, televisions, VCRs, telephones, etc. have HCI's. Common to these HCI's is the fact that the information is entered or retrieved in a specific format and sequence. The sequence of events by which a human and a computer communicate is known as a dialogue.
Dialogues occur in HCI's due to the fact that many decisions must be made. Breaks in the dialogue must occur at decision points. Once the task of determining the functionality that the system will possess through the Analysis and Specification phase is complete, the required dialogue must be mapped out. This task is best done visually via a flowchart. Figure 2.0 shows a simple Interface Structure Diagram for a library system.
Figure 2.0 - Interface Structure Diagram
In the diagram, the rectangles are nodes, which represent decision points or key entry or exit points in the dialogue. The arrows represent state transitions when the dialogue switches from one participant to the other. This can be the pressing of the Enter key, the clicking of a mouse button etc. They usually coincide with the completion of a specific function.
Research has shown that the best dialogues are created with the input from the actual user. The best designs are collaborative efforts between the HCI designer and the user. Prototyping is a key tool in this collaboration. If the user can actually sit at a terminal and use the dialogue, errors or improvements can quickly be found, faster than by the designer only who, in most cases, is not going to be the end user.
At this stage in the process, the basic structure of the dialogue is known. Even after the product is delivered to the end user, feedback on the HCI should be retrieved based on use and this feedback used to improve future versions. The next steps must be to begin to detail what occurs at the dialogue nodes. This deals with the specifics of how information is entered into the computer and how it is presented as output.
This section deals with determining what information should be used at what stages of the HCI. Presentation design deals with general principles of designing both input and output.
After the task of mapping the interface is complete, a number of input and output points in the process must be determined. These points coincide with events in the dialogue where information must be received by one party from the other in order to proceed. In general, the steps for determining these points and dealing with them are as follows:
1. Identify all system inputs and outputs.
2. Segment the dialogue specification into screens, overlays, windows, using closure events, such as the hitting of a key or the clicking of a mouse button, to determine boundaries.
3. Identify user requirements and characteristics. Before designing your display, know your user. This will help in deciding how much help to give, how many prompts etc.
4. Describe in detail the data that will be on each logical screen.
5. Design screen structure beginning with a general layout and adding additional information as required.
6. Test the screens with the users. Get their feedback and modify if necessary.
As a result of these steps it is possible to develop a concrete implementation of the required dialogue.
Data that is on a screen can be from several different categories.
1. Basic identification data must be part of every screen. What is the application? What is the current screen? Who is the current user? 2.Application data is dependent upon the current functionality. What data is currently displayed? What data is required input?
3. Support data is intended to aid the user. This comes in the form of help screens, pulldown lists, prompts etc., all intended to make sure that the user is never in a position of not knowing what to do or what can be done.
4.Navigation data deals with questions like: How do you leave the application? How can you undo changes? How can you advance to the next screen?
Also, the designer has to think about errors or extraordinary occurrences. What happens when the user enters invalid data? What can the user do when they aren't sure what to do? These are the types of questions that must be asked at this stage in the HCI design and development.
DATA ENTRY INERFACES
HCI dialogues require that two parties, namely the Human and the Computer communicate. At this stage the future dialogue has been determined and the method by which this dialogue will be implemented has been designed. Once it is determined what inputs are desired when, it remains to implement those input requirements. This section deals strictly with the input function. It mainly deals with the options the designer has and what best practices have been developed that he can utilize?
Many hardware devices exist for data input. Computers have keyboards and mice for data input, microphones for voice input, scanner for graphics input if desired, modem for input from fax, phone, and the Ethernet card, for input from the Internet, cable television, DVD etc. Other devices may include bar code readers, which are commonly used today in supermarkets, libraries, etc.
Primary methods of input are still the keyboard, mouse and increasingly, the microphone. The other devices listed tend to be secondary input devices in that the computer usually accesses them as the result of an instruction originally through one of the primary devices. Bar code readers can be the primary input devices, which initiate transactions on devices such as Automated Bank Tellers etc.
The keyboard is used to enter data into specific areas on a screen as the result of a response required from the computer. A key aspect of HCI is that the user knows what is required of him in regards to that response. This is done during screen design. As stated above, we know what the screen flow will be and what data we want on or from a specific screen. Improvements can be done to the screen to remove ambiguity and assist the user in understanding what is required from him.
Some guidelines that are commonly used to improve the screen layout and usage:
1. Set defaults for commonly entered items.
2. If codes and abbreviations are used, ensure that translations and interpretations are available.
3. Automatically fill in previously entered fields or data that can be retrieved from a file.
4. Use selection or pulldown lists if selections are from a previously determined set of options.
5. If the input screen resembles an already used and accepted form, try to use the form as a model.
6. Group like data. When data has a logical connection, place them on the same screen if possible without crowding. For example, all of a customerâ„¢s address information should be entered on a single screen.
A screen consists of fields that, when entered, are processed. Several types of processing are done, namely unit validation such as confirming that numeric values are only numbers, dates are valid etc., referential validation such as confirming that an entered customer number exists etc. Decisions must be made as to when the different types of processing takes place. Some general guidelines that are followed are:
1. Explicit Enter. Referential validation only occurs when the user presses the enter key. Unit validation can occur when the focus leaves a specific field.
2. Explicit Movement. If the user can have choices as to which field to enter next, force the user of the TAB or CR entries or reposition with the mouse.
3. Explicit Cancel. If the user interrupts data input, do not erase it but leave it visible for consideration. Only upon validation should invalid data be removed.
4. Explicit Delete. On any undoable transactions, force a confirmation. The user should also be warned of the undoability of the transaction so as to minimize any user error.
5. Feedback. Make sure that the user can see what they've entered.
6. Allow editing. While maintaining referential integrity, allow the user the ability to go back to previous data and modify.
7. Provide Undo. As far as possible, allow the user to undo previously entered transactions or actions.
8. Autoformat. Where there is a predefined required format, show that format on the screen or give the user the ability to determine that format.
9. Show valid responses. When the input is from a predetermined list of entries, allow the user access to that list either through the use of a pulldown list, selection buttons etc.
10. Let the user control. Let the user determine the rate of data entry. The designer cannot be assured that the application can handle all eventualities.
When designing a screen, it is essential to understand that a single screen consists of a number of different components. The following is a description of the user interface of Microsoft Word, which is used commonly for text processing. The top is the title bar that describes what the current application is. Beneath that is the menu bar which describes what the user can do. This consists of various options like File, Edit, Insert, Format, etc. Different options are clearly spelled out with a comprehensive help function available. Beneath that are tool bars that vary depending on what the user can do on the current screen. Where options must be chosen from a predefined list, such as for the font size and font name, a pulldown list gives all the options. Although this Word Processor has a relatively simple HCI it is sophisticated in the fact that the user controls the flow, has complete knowledge of what is expected and what options are available. Its basic features should be utilized in developing more complex applications.
DATA RETRIEVAL INTERFACES
Information can be retrieved from a computer in many ways. It is possible to read the output either on a screen or from a paper copy. It is also possible to have a voice synthesizer which will output the data through a speaker if the reader is vision impaired.
Other methods for retrieving information are: printers, plotters, signals to devices to start a device, stop a device etc., writing to storage devices. HCI deals mainly with screens, printed material, and audio. Similar principles apply to both screen and printed output as they are both visual. In other words, visual and audio are the two main HCI interfaces. The section on human physiology and psychology examined the capabilities and limitations of the human visual and audio senses. A successful HCI must adapt to these. The following guidelines should be considered when designing the Computer to Human component of an HCI.
1. Group data in a logical manner. Data that can be associated and has a logical relationship should be easily obtainable from the users current position within the application.
2. Order data according to meaning. People read from left to right. The most important information should be at the top left of a screen or report.
3. Use lists. When the relationships and similarities between various data can be seen all within the same field of view the relationships can be assimilated easier.
4. Use appropriate fonts and sizes. The user should not have to search a screen to find data because it's too small.
5. Provide clear headings, titles and navigational information.
6. Use colour to highlight important data. This can be error warnings or any other information that the application wishes to draw the usersâ„¢ attention to.
7. Use popup and dialog boxes for extraordinary situations. These can include errors but also one time date entries, logons etc.
8. Where there are different conventions, show the convention currently in use. For example, the conventions used for expressing the date vary between places. If the expected date input format is YYYY/MM/DD, make sure that the user knows this.
When the designer has a chance to use graphics, they should take that opportunity. It is especially true with HCI that a picture is worth a thousand words. Data that is spatial in nature should be shown on a map. Graphs or pie charts should be used to represent statistical data. Ensure that any graphs are well labeled. All axes should be labeled with legends for all displays. It should also be ensured the scale is known. All key points should be labeled so that anyone using the graph can clearly see what is being represented.
The software engineering process is followed for the design of an HCI. As with any stage in systems design, the user must be involved early in the design process. Mockups and prototypes have clearly been proven as effective tools to improve HCI design and the sooner the end user gets a chance to comment on the product, the better the product will be.
This seminar and presentation attempts to give an introduction to the subject of designing and developing good Human Computer Interfaces. Understanding how the human perceives information and its capabilities and limitations is of prime importance in developing workable HCI's. Several proven methods were explored which should assist the HCI designer in determining what interfaces have to be developed and how to make them so that they assist the user in accomplishing the desired tasks.
The techniques that can be used from a computer perspective are changing quite rapidly. The fact that this field has undergone and will go through a lot of changes is reflected in the way computer technology has developed. Only some years back an HCI involved punch cards, paper tape and massive printouts. Today terminals, keyboards, mice, speakers, touchpads etc take up the main role. Twenty years from now it is possible to have highly advanced forms of 3D virtual reality displays and voice interpreters, all which are on the market now in a primitive way. Recent technology that enables transmission of the sense of touch is only an indication of the things to come. HCI will be an ongoing area of research for many years to come.
Â¢ Trends in Software Engineering â€œ The Human Computer Interface- Alan D. Goodbrand
1. INTRODUCTION 01
2. HUMAN PHYSIOLOGY & PSYCHOLOGY 02
3. INTERFACE ANALYSIS &SPECIFICATION 08
4. DIALOGUE DESIGN 12
5. PRESENTATION DESIGN 14
6. DATA ENTRY INTERFACES 16
7. DATA RETRIEVAL INTERFACES 20
8. CONCLUSIONS 22
9. REFERENCES 23
Praise be to God.
I express my sincere thanks to Prof. M.N Agnisarman Namboothiri (Head of the Department, Computer Science and Engineering, MESCE), Mr. Sminesh (Staff incharge) for their kind co-operation for presenting the seminar and presentation.
I also extend my sincere thanks to all other members of the faculty of Computer Science and Engineering Department and my friends for their co-operation and encouragement.
FAIS JAHAN MUSTAFA
Use Search at http://topicideas.net/search.php wisely To Get Information About Project Topic and Seminar ideas with report/source code along pdf and ppt presenaion
Active In SP
Joined: Feb 2011
18-03-2011, 11:59 AM
human cmp interface2.doc (Size: 71.5 KB / Downloads: 140)
Eyegaze Human-Computer Interface
The Eyegaze System provides an eye-controlled human-computer interface (HCI), allowing people to interact with computers by pointing with their eyes. A video camera mounted below the computer monitor unobtrusively observes the user's eye and specialized image processing software analyzes the video images of the eye and determines the eye's gaze point on the monitor screen in real time. Early applications of the Eyegaze System addressed an HCI for people with severe motor disabilities. Simply by looking at control keys displayed on a computer screen a disabled user can type, generate synthesized speech, control lights and appliances, operate a telephone, play games, and run DOS-compatible off-the-shelf software.
Creating an eyegaze HCI that accommodates a variety of physical disabilities presented our engineering team with several technical challenges. The Eyegaze System has to be accurate enough for a user to trigger the 5/8-inch keys of an on-screen computer keyboard. The calibration procedure needs to be simple, and the system needs to maintain calibration when the user leaves the computer and returns. Finally, for people with uncontrolled head motion, it needs to be tolerant to head motion. The accuracy and calibration objectives have been achieved. A solution for accommodating head motion is under development.
There are many kinds of eye tracking devices, ranging from galvanometric sensors which measure voltages across the eye, to video image processors which examine optical images of the eye (Mason, 1969; Merchant and Morrisette, 1973; Cornsweet, 1973). Eye trackers employing image processing are by far the most accurate and reliable, and are therefore preferable (Young and Sheena, 1975).
Image processing eye trackers exist in two categories: head-mounted and remote. For disabled people operating computers, it is appropriate to sense the eye unobtrusively, with remotely mounted cameras. The user need not be mechanically "hooked up" to access the system, and has no need for cumbersome equipment on his body.
In 1988, LC Technologies completed development of the first Eyegaze Computer System designed for use by people with severe motor disabilities. Eyegaze is a PC-based system, requiring only the control of one eye. Selections are made by fixing the gaze in control "keys" on the screen. Nothing is attached to the user.
As illustrated in Figure 1, a video camera located below the computer screen continually observes the user's eye, and specialized image-processing software determines the eye orientation and project and implimentations the subject's gazepoint on the computer display. With a person sitting between 18 and 24 inches from the computer screen, the system predicts the gazepoint with an average accuracy of better than 1/4 inch. The system also generates information regarding pupil diameter, blinking, and eye fixations, useful for other eyetracking applications.
The Eyegaze System uses the pupil-center/corneal-reflection method to determine the eye's gaze direction. A low-power infrared light emitting diode (LED) located in the center of the camera lens illuminates the eye (Hutchinson, 1989). As shown in Figure 2, the LED generates a small, very bright reflection off the surface of the eye's cornea and, because it is located at the center of the camera lens, the LED causes the bright-pupil effect by reflecting light off the retina. The computer calculates the person's gazepoint, i.e. the coordinates of where on the display he is looking, based on the relative positions of the pupil center and corneal reflection within the video image of the eye.
Prior to operating the eyetracking applications, the Eyegaze System must learn several physiological properties of a person's eye in order to be able to project and implimentation his gazepoint accurately. It must know the radius of curvature of the eye's cornea and the angular offset between the eye's optic and focal axes. The system learns these parameters by performing a calibration procedure. To calibrate, the user fixes his gaze on a sequence of small circles that the computer displays at specific locations on the screen. The calibration procedure usually takes about 15 seconds and can be performed independently.
The Eyegaze System can save calibration results for future use, and it will retain current calibration data even if the user moves away from the system. When he returns to his position in front of the camera, Eyegaze will resume its gazepoint determination, enabling the user to continue to operate the system
Active In SP
Joined: Feb 2011
19-03-2011, 02:59 PM
SANJAYA KUMAR JENA
hci.pptx (Size: 104.21 KB / Downloads: 143)
The study of the design, evaluation, and implementation of interactive computing systems for human use and the major phenomena surrounding them.HCI often called as sociotechnological discipline.
Three parts: the person, the computer, andBy which they work together
the others in the organisation
user “tells” the computer what (s)he wantscomputer communicates results
HCI draws the knowledge from computer science ,graphic design, management, psychology and sociology etc..
By using this one can able to learn foreign language, analyzing aerodynamics of new aero plane, planning surgery, playing computer game, accessing information on world wide web.
GOALS OF HCI
Design and develop systems that are usable, efficient, and safe.
Design and develop systems that are intuitive.
Allow people to use them with a minimum of change and disruption.
Make data exchange between people and machines less stressful and less chance of Misunderstandings.
USABILITY IN DESIGN
Usability is one of many factors that software designers must take into account when designing a
New user interface.
Usability sometimes neglected leads to greater cost
A skilled interface designer can’t miss the important information.
COMPONENTS OF USABILITY
DESIGN OF HCI
REQUIREMENTS ANALYSIS AND SPECIFICATION
2 unit testing
INTEGRATION AND SYSTEM TESTING
EXCELLENCE OF HCI
Quality of life.
Growth of computer and communication industries.
Joined: Jul 2011
27-01-2012, 10:49 AM
to get information about the topic eye gaze human computer interface full report,ppt, related topic refer the link bellow
Joined: Apr 2012
23-04-2012, 02:59 PM
Human Computer Interface HCI
hci.pdf (Size: 7.59 MB / Downloads: 68)
Zeiss Slide Projector
• Only one button to control the slide advance, see Figure 2.3.
• During lectures, sometimes the slides go forwards, sometimes they go backwards . . .
• If you can find an instruction manual:
Short press = forward, long press = backward.
• What an elegant design, two functions with just one button!
• But how should first-time users know what to do?
The Louis-Laird Amphitheatre in the Sorbonne
• Magnificent murals on the ceiling.
– But only the right way up for the lecturer.
• Electric project and implimentationion screen.
– Has to be lowered from a back room up a short flight of stairs, out of sight.
• The audiovisual control trolley (see Figure 2.5) at the front of lecture theatre HS EDV also caused
me a major problem.
• Intending to hook up my laptop to the ceiling mounted project and implimentationor, I unwittingly unscrewed the
monitor cable connector from the outside of the trolley.
• This resulted in a dull clunking sound from inside the trolley.
• Unfortunately, the internal connector had only been secured by virtue of its being attached to the
Where is the Toilet Paper?
• Fancy hotel, nice bathrooms, see Figure 2.7.
• Having sat down and done the business, where the heck is the toilet paper?
• Ah, there it is! Well-hidden, see Figure 2.8.
• Shower control: water either goes into the bath out of the faucet or comes out of the shower. See
• Sticker with instructions on the faucet.
• How do you make the water come out of the shower instead of the faucet?
• You have to reach under the faucet and pull the knob down!
The Icon Design Lifecycle
Usability engineering lifecycle for icons:
Design, test, redesign.
Icon Design Iterations
• Start with simple black and white, hand-drawn sketches on paper (silhouette conveys the most
• Test and redesign until the basic symbols work.
• Add greys and perhaps colour. Design on computer. Print out colour versions of the designs, at
the approximate real size.
• Test and redesign until the icons work.
Icon Intuitiveness Testing
Test the intuitiveness of (a set of) icons by running a simple thinking aloud usability test:
• Mount each icon design on a piece of card.
• Either tape up an area of the table with masking tape, into which the icons are placed,
This is so the video camera remains focussed and test users are less
tempted to pick up the icon cards.
Better still, use a stand to present icon designs at approximately the correct viewing angle and
Joined: Apr 2012
23-05-2012, 11:32 AM
Human Computer Interface
Human Computer Interface.ppt (Size: 784 KB / Downloads: 48)
Understand the need to facilitate an effective dialogue between humans and machines.
Explain the need to design systems which are appropriate to users at all levels and in different environments, e.g. the impact of clarity of structure and layout.
Describe how the user-interface can be designed for effective communication with the user.
Describe the advantages of common user interfaces between different generic application packages.
Describe the advantages and limitations of a natural language interface.
What is an Interface?
A point where two objects meet.
A point where the human can tell the computer what to do
A point where the computer displays the requested information
What Tools are Used?
A keyboard, for typing,
A mouse, for clicking,
A scanner, for copying,
A camera, for images.
A monitor, for displaying,
A printer, for printing,
A sound card. For audio,
A DVD, for video.
A good interface match would include as many senses as possible.
Computer input and output is basically seeing what we enter and what is displayed.
Sound can be added to some programs, either by giving instructions by voice, or listening to a commentary / music.
Touch can be in the form of using the mouse, a joystick, or a drawing tablet.
There are many featured designed for disabled users.
A visually impaired person may be unable to use the GUI features, and find that typing in commands are easier. The printed output may be produced on Braille.
A hearing impaired person may be able to use speech commands and listen to the output through speakers. This may also be a preferred method for users with other disabilities who cannot use a keyboard and mouse.