Read the TechNews Online at: http://technews.acm.org
ACM TechNews
May 31, 2006

MemberNet
The ACM Professional Development Centre
Unsubscribe

Welcome to the May 31, 2006 edition of ACM TechNews, providing timely information for IT professionals three times a week.

Sponsored by Information, Inc.

http://www.infoinc.com/sponsorthenews/contactus.html


HEADLINES AT A GLANCE:

 

RAD Lab Wins Backing From Five Major IT Firms
UC Berkeley News (05/30/06) Moore, Theresa

The University of California, Berkeley, has announced that IBM, Hewlett-Packard, NTT, Nortel, and Oracle are partnering with the university's Reliable, Adaptive, and Distributed (RAD) systems laboratory (RAD Lab), with each pledging up to $170,000 in annual contributions over the next five years. The partnership illustrates both the importance of the RAD Lab's work on developing the next generation of Internet design tools and the changing dynamic of long-term university research funding. "Until recently, federal grants from the Department of Defense's Defense Advanced Research Projects Agency and the National Science Foundation covered research costs for large projects such as this," said ACM President David Patterson, founding director of the RAD Lab. "But reductions in government funding of information technology research led us to seek alternative funding sources." The RAD Lab is trying to adapt new discoveries in machine learning to manage the vast, distributed computing systems required by data-intensive Internet companies. The lab is developing software to automate the time-consuming tasks of creating and maintaining these systems. "System failures and security breaches are facts of life, so the most practical solution is to focus on fast detection and recovery from failures, as well as on immediate detection and isolation or containment of security breaches," Patterson said. "We hope to automate that with statistical machine learning." Any code produced by the RAD Lab will be freely available to the public under the Berkeley Software Distribution License. The sponsor companies will dispatch consultants to advise the center's activities, though none will actually work at the lab. Patterson says the companies will benefit from observing the development process, and will be able to guide the researchers by presenting obstacles that could arise when the technology is put to commercial use.
Click Here to View Full Article
to the top


Debating the Bugs of High-Tech Voting
Washington Post (05/30/06) P. A15; Goldfarb, Zachary A.

As midterm elections near, the debate centering on the security of e-voting systems is heating up as voting machine vendors and voting rights activists clash over the severity of a recently discovered vulnerability. The vulnerability, discovered in a Diebold machine several weeks ago in Utah, would enable anyone with a rudimentary knowledge of computer programming to manipulate the code and alter votes in just a few minutes time, security researchers claim. California and Pennsylvania issued a warning to all counties in those states that use the Diebold machines, though the degree of the threat remains a point of sharp disagreement. David Jefferson, a computer scientist at Lawrence Livermore National Laboratory, was shocked when he found out about the threat exposed in Utah, and echoed the "frequently expressed opinion that this is the worst vulnerability we have ever seen." Diebold counters that the vulnerability was a product of design, that it is there to enable the machines to easily receive software upgrades, and that a person could only tamper with election results if given unrestricted access to the machines. E-voting systems came into widespread use after the Help America Vote Act (HAVA) of 2002, which was passed in the wake of the disastrous 2000 presidential election, but electronic machines do not improve the reliability of voting process without a manual paper trail, voting rights advocates argue. While it has never been proven that ballots have been manipulated in an actual election, numerous votes have been delayed by flaws in the technology. The federal Election Assistance Commission, created to assist in HAVA implementation at the state level, claims that it is in the process of improving the election-system certification process, though voting-rights groups in several states have been pursuing legal action to halt the purchase of new electronic systems. For information on ACM's e-voting activities, visit http://www.acm.org/usacm.
Click Here to View Full Article
to the top


Senate Immigration Bill Raises H-1B Limit
IDG News Service (05/30/06) Gross, Grant

Tucked within the Senate immigration reform bill that has become the focal point of a contentious national debate is a provision calling for an expansion of the number of H-1B visas allotted annually to highly skilled foreign workers. Passed by the Senate on Thursday, the bill would raise the annual H-1B cap from 65,000 to 115,000, but it has come under fire from House Republicans who claim that it takes a soft stance on illegal immigration. Technology companies have been calling for more H-1B visas for some time, claiming that it is becoming increasingly difficult to find qualified U.S. workers. By passing the bill, Microsoft Chairman Bill Gates credited the Senate with having taken a "critical step forward in its important work to ensure that our nation remains the global leader in technology innovation." A group working on behalf of U.S. IT workers claims that the H-1B program is fraught with abuses, and that companies pay H-1B workers below their market value. "The program is basically broken and can easily be manipulated," said Ron Hira of the Institute of Electrical and Electronics Engineers. "Until it's fixed, it makes no sense to increase the cap." Despite clearing the Senate, the bill faces stiff opposition from Republican lawmakers and an uncertain future.
Click Here to View Full Article
to the top


Outward Bound for Robots
Technology Review (05/31/06) Graham-Rowe, Duncan

Scientists have tested navigation software based on the hippocampus of the human brain in a self-driven robotic car. The system allows the robot to take "cognitive fingerprints" of its environment so that it can explore new places while building a living memory. In testing, the robotic vehicle, complete with a laser range finder and omnidirectional camera, successfully navigated more than one and a half kilometers in an urban environment before it got lost. The system has also been put to the kidnapping test, where an indoor robot is "blindfolded," taken to an unfamiliar location, and left to then find its way home. The kidnapping test raises the problem of simultaneous localization and mapping (SLAM), an area of increasing importance as robots begin to take on more autonomous functionality. The difficulty arises when trying to create a map that will help a robot navigate an area that it is still exploring, which requires the robot to map foreign terrain while simultaneously updating its own position. "To localize the robot, a map is necessary, and to update a map, the position of the mobile robot is needed," said Adriana Tapus, a roboticist at the University of Southern California who designed the system. The sensor measurements also carry a degree of inherent uncertainty, which can throw off the accuracy of the maps. Tapus' technique mimics the way that people navigate by culling together raw data from the robot's sensors, such as corners and colors, and combining them into a basic description, or fingerprint, about the location. Tapus maintains that it is the combination of the features that makes the fingerprints unique, which she believes is the same technique used by humans to remember landmarks. Scientists have identified the distinct firing patterns of the cells in a mammal's hippocampus that are responsible the animal's memory of locations, recreating the technique with probabilistic algorithms.
Click Here to View Full Article
to the top


MSU Professor to Try New Teaching Method in Germany
Montana State University (05/31/06) Ellig, Tracy

Montana State University computer science professor John Paxton plans to use problems from programming competitions to teach students about data structures and algorithms. Paxton, who has served as an MSU advisor for the Association for Computing Machinery (ACM) for 16 years, will try the new teaching method on students at the University of Leipzig in Germany, where he is spending the 2006-2007 academic year after receiving a Fulbright Scholar Award. Students are typically shown which data structures and algorithms will solve a particular problem, but they do not gain a greater appreciation for the complexity of the solutions. "The traditional way of teaching this topic takes the excitement and challenge out of the material," says Paxton. He has seen firsthand how exciting the subject matter can be for students involved in programming competitions such as the annual ACM event, in which teams of three students have five hours to solve eight problems on a single computer. "They come back very motivated to learn more computer science because the competition points out what knowledge they're missing," says Paxton. The response of Leipzig students to competition problems will determine how much of the course he will incorporate into his teaching methods upon his return to MSU.
Click Here to View Full Article
to the top


Incidents Prompt New Scrutiny of Airplane Software Glitches
Wall Street Journal (05/30/06) P. A1; Michaels, Daniel; Pasztor, Andy

As commercial airplanes grow more dependent on increasingly complex computer code, software glitches are emerging as a primary safety concern. The systems in the latest jetliner contain more than 5 million lines of computer code, compared to fewer than 1 million in older models, making it increasingly difficult to locate the flaw when something goes wrong. While the software used in planes is tested far more rigorously than everyday office applications, the errors that inevitably arise can have much more serious consequences, and officials have begun reviewing flight data from earlier accidents to determine what role, if any, faulty software played. "It's our next big area of work," said Peggy Gillian of the FAA, adding that only recently officials and experts "came to the realization that we haven't looked at this area" closely enough. While no airplane crash has yet been attributed to malfunctioning software, several recent incidents where software glitches have disrupted flights have called attention to the problem. "A total loss of flight control could be worse than a fire on board," said Robin McCall, a retired pilot for Delta Air Lines. McCall says that automation programs can make it difficult for pilots to revert to manual controls to overcome a problem. Experts agree, however, that air travel has become much safer since the introduction of automated systems. Today's autopilot systems handle much more of the plane's functions than when they were first designed, including adjusting the cabin's air pressure, optimizing fuel efficiency, and warning of the threat of collision or mechanical breakdown. Forthcoming models, such as Boeing's 787 Dreamliner, will bring new levels of automation, replacing autonomous hardware and software with redundant central computers. A group of U.S. airlines, pilots' unions, jet manufacturers, and software vendors recently launched a data-gathering initiative to analyze previous computer-related accidents.
Click Here to View Full Article - Web Link to Publication Homepage
to the top


Intelligent Beings in Space!
New York Times (05/30/06) P. D1; Chang, Kenneth

The increasing distance and duration of space exploration missions has led scientists to develop new and faster methods of interplanetary communication. Scientists are using artificial intelligence techniques to automate spacecraft, looking to devices such as iRobot's robotic vacuum cleaner Roomba for inspiration. Intelligent software already manages the schedule of Earth Observing-1 (EO-1), a satellite that monitors the Earth for events such as volcanic eruptions, scanning for changes that could indicate a natural disaster. "Almost immediately, within a matter of hours, the spacecraft is reprogramming itself to image these targets," said NASA's Steve Chien, "and we can get a very rapid response imagery of breaking science events." The computer can readjust the satellite's schedule, postponing other observations based on the data it collects and the data supplied by sensors and other satellites. The planning software pruned EO-1's budget by one-quarter when it was introduced in 2003, and it could see future use in new planetary explorations, such as the next mission to Jupiter and its moons. NASA's two Mars rovers both contain stereo cameras that enable them to navigate around rocks and other obstacles, and they will soon receive an upgrade to speed the process of relaying instructions from mission control. Another software improvement will enable the rovers to take a preliminary look at their data and only send back the images that are useful. Tests have shown that the software correctly identifies the clouds that the rovers are studying 93 percent of the time, which will dispense with the time-consuming transfer of meaningless data. The software still has a long way to go, however. "None of the AI systems are as smart as a 2-year-old," said NASA's Cynthia Cheung. New applications will incorporate machine learning, where a computer's intelligence builds based on its collected experiences.
Click Here to View Full Article - Web Link May Require Free Registration
to the top


Keeping Computers in Check
CNet (05/26/06) Terdiman, Daniel

If left to develop unchecked, computers could become so dependent on algorithms that they could overlook basic issues of livability as they execute tasks such as determining the maximum number of buildings in a neighborhood, warns Sheldon Brown, the head of the University of California at San Diego's Center for Research in Computing and the Arts, where he develops and oversees research on the power of algorithm-based content. "Just because we can do something, doesn't mean we should," Brown said. "When we build tools, we have to think deeper and broader about what the tool is doing." Brown has developed a preliminary version of a software application called "Scalable City" to help researchers explore the extent to which computational power can aid and even automate urban planning and video game development. Using "L-curves," the multimedia Scalable City can digitally fill the image of a barren landscape with rows of new houses flanking elegantly curved roads. Brown warns that while the neighborhood looks ideal from an aerial view, on the street at the micro level, the computer cannot know how livable the environment will actually be for humans. The tendency of the algorithm-based Scalable City to design neighborhoods too crowded with buildings for human inhabitants illustrates Brown's fear that looking at computer power from the macro perspective ignores the logical errors that are only visible on the micro scale. Scalable City could also lead to practical applications in video game design within a couple years, such as new tools for enabling multiple players to control and engage in the games. Brown notes that humans in many ways already constitute a living data set. "Our lives in general are now really a part of algorithmic processes," he said, adding that the "major part of our lives are for producing algorithmic data and then interacting with it. We've become I/O, input/output."
Click Here to View Full Article
to the top


The Next Wave of the Web
Nature (05/26/06) Butler, Declan

In a recent interview, Nigel Shadbolt, chair of the panel 'The Next Wave of the Web' at the recent WWW2006 conference, discussed the current and future states of the Web. The dominant theme at the conference was the emergence of wireless broadband for mobile devices, which will bring the concept of ubiquitous computing closer to reality than ever. The emergence of pervasive mobile computing will bring the Web to millions of people around the world who cannot afford computers or live in areas with no network infrastructure. The conference also focused on Web 2.0 in the context of solving problems through the broad participation of many users. Shadbolt noted computer users' growing willingness to share data on the Web, pointing to the mash-up phenomenon as evidence of the potential of the open-data movement. Microsoft's Tony Hey gave a presentation on e-science, or the trend of scientists using massive, distributed datasets to conduct research. Shadbolt, whose background is in artificial intelligence, believes that the Web will be the environment where the discipline will finally realize its potential, noting that the Bayesian statistical methods that enable machines to make decisions already pervade the Web, such as the open-source Bayesian spam blocker, SpamBayes. Shadbolt is also optimistic that the Semantic Web could finally begin living up to its expectations. Even simple descriptions about the meaning of data, such as RDF, could greatly improve the functionality of the Semantic Web, Shadbolt says, noting that a basic RDF tag can describe key data classes and relationships. Shadbolt believes that the next major developments will be in displays, with flexible screens and laser projection fast becoming commercially viable.
Click Here to View Full Article
to the top


Intelligent Machines
UC Irvine New University (05/30/06) Kim, Jocelyn

By storing human cognition for all eternity, intelligent technology could bring an end to human mortality, according to Marvin Minsky, a professor of electrical engineering and computer science at MIT who recently gave a speech on the future of intelligent machines. Intelligent machines could also safeguard human mortality by conducting dangerous missions, though Minsky notes that most robots in production today are designed for trivial tasks. "All over the world, people are building the same robots. People should be working on something more useful," he said. "Could we make a clever machine other than those that play soccer?" Minsky believes that researchers should concentrate on analogy-based problem solving, rather than neuron networks, genetic algorithms, and other applications that he dismisses as "fads." He attributes machines' inability to understand language to the lack of sufficient programming. "The reason why computers don't understand words is no one has made a semantic dictionary," Minsky said. "Common words have a dozen meanings. You need to have representations of different things and links between similar definitions."
Click Here to View Full Article
to the top


Image Binarisation
Nation (Thailand) (05/26/06) Sutharoj, Pongpen

The best computer science project at the Intel International Science and Engineering Fair (Intel ISEF 2006) went to Natt Piyapramote, a 17-year-old student from Thailand. Piyapramote won for his automated image binarisation software, which will allow users to bypass the use of an optical character recognition (OCR) scanner in converting document images from conventional digital cameras to a black-and-white image file. "Just put the image file into a computer and have the software proceed," says Piyapramote. "The outcome is automatic with no manual involvement." Piyapramote's statistical-based Adaptive Binarisation for Document Imaging project, which has also won an award from ACM, has an acceptable 14 percent error rate, and although it is just 5 percent off the best in the world in terms of efficiency, it offers faster processing and is more automated. He plans to lower the error rate to 5 percent, and put the software on the Web free of charge for those who want to simplify the conversion of document image files into text files in order to edit information. The Young Scientist Competition participant also won an award from the American Association for Artificial Intelligence for the best project with an artificial intelligence component.
Click Here to View Full Article
to the top


Codes on Sites 'Captcha' Anger of Web Users
Wall Street Journal (05/31/06) P. B1; Kesmodel, David

Web sites such as Yahoo.com protect themselves from mischief-making programs by having users who wish to gain access solve puzzles known as "captchas," which often take the form of a visually distorted code that must be correctly typed. Keeping pace with new spamming techniques and strategies has prompted some Web sites to make these codes trickier to solve, irritating more and more Web users. "We know there's no perfect panacea, but we think this is a great tool to prevent malicious activity," says Google engineering director David Jeske. However, the World Wide Web Consortium published a paper in November 2005 warning that captchas "fail to properly recognize users with disabilities as human" and can be thwarted by clever programmers, as part of its argument for programmers to develop alternative captchas. The group noted, for example, that spam companies sometimes use people rather than automated programs or "bots" to decipher the captchas. Director of the consortium's Web Accessibility Initiative Judy Brewer says visual captchas are especially difficult for disabled people because they "don't tell humans and computers apart; instead, they tell able-bodied humans and computers, along with disabled humans, apart." Alternatives some Web sites are exploring or deploying include audio captchas or quizzes that involve simple problem-solving. In development by Lehigh University computer science professor Henry Baird are "scattertype" captchas that fragment each letter in the code, while some sites are simplifying their captchas so humans can solve them easier.
Click Here to View Full Article - Web Link to Publication Homepage
to the top


IBM Looks to Appalachian Colleges for Future IT Professionals
Computerworld (05/26/06) Weiss, Todd

As part of its educational partnership with the Appalachian College Association (ACA), IBM, through its two-year-old Academic Initiative program, will provide the association's 35 two- and four-year liberal arts colleges and universities with $5 million in software, technical services, and discounts on software in an effort to boost the number of students studying information technology. IBM held a workshop last week at Bryan College in Dayton, Tenn., that enabled professors from ACA schools to train on the company's donated technology so that they can better prepare their students for careers in IT. By next fall, about 350 to 500 students from ACA colleges and universities in Kentucky, North Carolina, Tennessee, Virginia, and West Virginia are expected to enroll in classes that will be taught by professors who have attended the IT workshops and have been trained on various technologies. "IBM is concerned about where its next IT workers are going to come from," says Martin Ramsey, chief instructional technologist for the nonprofit ACA consortium. IBM also believes the region has the potential to develop some high-tech startups. "A lot of the next Googles could come out of central Appalachia," says Mark Hanny, vice president of the Academic Initiative.
Click Here to View Full Article
to the top


The Next Big Bang: Man Meets Machine
CNet (05/29/06)

The fusion of digital technology and human form could become a reality as scientists continue to make discoveries in biotechnology and nanotechnology. Scientists are also looking to the biological world to replace some mechanical parts of a computer with proteins, DNA, viruses, and bacteria. The symbiotic confluence of different these strains of scientific endeavor is known as "BANG," short for bits, atoms, neurons, and genes. "All these things are converging because biology, nanotech, and organic chemistry are running together," said Mark Bunger of Lux Research. When 24-year-old Matthew Nagle was stabbed in the neck in July 2001, the blow severed his spinal cord and he was instantly paralyzed from the neck down. Nagle volunteered for an experimental treatment that implanted a chip into the part of his brain that controls motor functions, detecting the electrical activity when the neurons fire that indicates "movement intention." The software reads the intention and translates it into instructions that manipulate a cursor on a screen. Scientists are also picking up on the technology to combat Lou Gehrig's Disease, or ALS, a disease much more complicated and diffuse than a spinal injury. Researchers are also turning to biology to improve the human-computer interface, particularly as silicon chips and other existing technologies reach their physical limitations. To that end, scientists have been exploring the basic building blocks of life--DNA, enzymes, proteins--as the basis for a repository of biological components that could help create synthetic organisms. As the volume of digital content rapidly proliferates, scientists are also exploring new storage technologies that resemble DNA. The military is also investigating biotechnologies to give paralysis sufferers greater autonomy and to enable able-bodied soldiers to operate equipment remotely.
Click Here to View Full Article
to the top


Why the Democratic Ethic of the World Wide Web May be About to End
New York Times (05/28/06) P. 9; Cohen, Adam

The effort by Internet service providers to impose a new system of fees on the Web poses a threat to Web creator Sir Tim Berners-Lee's vision of a platform on which everyone in the world could communicate on an equal basis, writes Adam Cohen. The new system of fees could create a tiered Internet that would enable service providers to shut out Web sites whose politics they do not agree with. Even if ISPs did not discriminate on the basis of content, access fees would automatically marginalize smaller, poorer sites. For example, Internet users can now watch video from content providers such as BBC World as well as video blogs and Web sites such as YouTube.com, where people can upload videos of their own creation. However, under tiered pricing, Internet users may be able to get videos only from major corporate channels. Berners-Lee, who has begun speaking out in favor of net neutrality, predicts that the fees could also hamper future innovations, such as a Web site that will allow Internet users to take videos of an event with their cell phones and piece them together to create a three-dimensional image of what happened. He also argues that service providers may be hurting themselves by pushing for tiered pricing, because customers who are used to the Web as it is now may not pay for access to a Web that is restricted to wealthy corporate content providers.
Click Here to View Full Article
to the top


Interactive Display System Knows Users by Touch
New Scientist (05/25/06) Sandhana, Lakshmi

Researchers at the Mitsubishi Electric Research Laboratories (MERL) have employed touch-sensitive technology in the development of an interactive computer display that is able to differentiate between multiple simultaneous users and keep track of them. "Most touch screens only permit one touch at a time," says Paul Dietz, a researcher at MERL in Massachusetts. "A much smaller number allow multiple, simultaneous touches, but none of these can tell you who is touching where." The DiamondTouch (DT) system is designed to assign a distinct electrical signal to individuals as they touch the surface of the screen, and send the signals through their bodies to the receiver located in their seat. The connected computer knows exactly where people have touched the screen, and distinguishes each user's touch all at once. The identity technology enables DT to track the input of each user and even limit access to certain functions. MERL researchers view the DT system as having potential applications as an addition to controls at power stations, the cockpit of an airplane, or the dashboards of cars and trucks. Clifton Forlines adds DT could also be used in gaming to create smarter multiplayer games.
Click Here to View Full Article
to the top


Eclipse Sets Off a Big Bang
SD Times (05/15/06)No. 150, P. 28; Burd, Barry

In late June the Eclipse Foundation will roll out major updates to 10 of its most prominent projects in the Callisto Simultaneous Release, whose goal is to guarantee that the projects can interoperate without interfering with each other, writes Barry Burd, a professor in the mathematics and computer science department at Drew University. Each project will remain a separate open-source initiative, according to an official Web site. "The Callisto effort will be more or less invisible to the user community, because the things that are going to happen as a result of Callisto are that the projects are going to get cleaner," says the Eclipse Foundation's Bjorn Freeman-Benson. Callisto seeks to make users' lives easier by requiring all project developers to offer enabling and disabling for their project's features, and by using the new Coordinated Update sites to obtain Eclipse plug-in downloads with little fuss. Callisto projects include the Eclipse Project update Eclipse 3.2, which features scalability, extensibility, enterprise readiness, ease of use, Java 6 compiler compliance, and a Java Code Clean Up wizard. Other initiatives focus on improving the user level experience, such as the Business Intelligence and Reporting Tools (BIRT) project to deliver the ability to generate business reports in Web and PDF formats. All Callisto projects will employ the Java version of IBM's International Components for Unicode libraries, or ICU4J. The next Eclipse update after Callisto may have a greater concentration on unifying the various projects, and the Callisto model's long-term hope is to instill another layer of respectability within open-source software.
Click Here to View Full Article
to the top


Teaching Reliability
IEEE Distributed Systems Online (05/06) Vol. 7, No. 5,Regehr, John

John Regehr of the University of Utah's School of Computing believes embedded systems education should prioritize software reliability. He assigns most of the blame for unreliable software not to students, but to the single-approach, "cowboy programming" strategy that students become used to doing because it works so well for small, well-defined projects in first- and second-year programming courses. Regehr lists partial and entertaining solutions that could solve the problem by enhancing embedded programming assignments. One solution is a code review immediately following the assignment's completion, augmented with class queries about design and implementation. Another suggestion of Regehr's is for students to present correctness arguments, which creates an incentive for the students to account for them during the design and implementation phase. "I devote a considerable fraction of my embedded software course to describing the available embedded software architectures, each of which comes with a very specific set of tradeoffs," notes Regehr. "The lesson I try to impart is that the choice of software architecture should be deliberate, rather than evolving out of a sequence of greedy choices." Regehr also stresses an emphasis on testing and interoperation, and reports that the sting of failure can be a valuable reminder of students' need to practice cautious development that emphasizes thinking ahead, assertions, testing, tools, and incremental deployment.
Click Here to View Full Article
to the top


A Review of National Information and Communication Technologies (ICT) and a Proposed National Electronic Initiative Framework (NEIF)
First Monday (05/06)No. 11,Peslak, Alan R.

The digital divide--the chasm between those to whom information and communications technologies (ICT) are easily available and those to whom they are not--is an undeniable fact and its continued existence reflects the lack of a national electronic initiative framework (NEIF) to improve access and utilization of ICT, writes Penn State University assistant professor of information sciences and technology Alan Peslak. He proposes such a framework as a tool that national and non-governmental organizations can use to bridge the digital divide. Peslak suggests that all national electronic initiatives be based on general doctrines built upon the five general rights of man espoused in the political philosophies of John Locke and Thomas Jefferson: Provision of equality, protection of life, acquisition and retention of liberty, pursuit of happiness, and protection of property. Electronic initiatives must then address specific directives that correspond with the five doctrines, namely universal access to ICT, societal enhancement, better electronic government, economic improvement, and legal and regulatory control. These directives can be applied to general domains, such as information access, communications access, and infrastructure for universal access; education, health, environment, agriculture, international issues, the incorporation of ICT into everyday life, information content, and law and order for societal enhancement; procurement, efficiency, access to government data, citizen-government communications, and government services for improved e-government; research and development and competitiveness for economic betterment; and intellectual property, security, cyber-crime, and privacy for legal and regulatory control. The final level in the NEIF consists of deliverables, the products and services furnished by the electronic initiative. Peslak concludes that the current state of the nation will by necessity reflect on the specific deliverables each domain addresses.
Click Here to View Full Article
to the top


To submit feedback about ACM TechNews, contact: [email protected]

To unsubscribe from the ACM TechNews Early Alert Service: Please send a separate email to [email protected] with the line

signoff technews

in the body of your message.

Please note that replying directly to this message does not automatically unsubscribe you from the TechNews list.

ACM may have a different email address on file for you, so if you're unable to "unsubscribe" yourself, please direct your request to: technews-request@ acm.org

We will remove your name from the TechNews list on your behalf.

For help with technical problems, including problems with leaving the list, please write to: [email protected]

to the top

News Abstracts © 2006 Information, Inc.


© 2006 ACM, Inc. All rights reserved. ACM Privacy Policy.