User:Waleed mufti2

TECHNOLOGY The Toronto Magnetic and Meteorological Observatory is a small observatory located on the grounds of the University of Toronto. The original building was constructed in 1840 as part of a worldwide research project run by Edward Sabine to determine the cause of fluctuations in magnetic declination. Measurements from the Toronto site demonstrated that sunspots were responsible for this effect on Earth's magnetic field. When this project ended in 1853, the observatory was taken over by the Canadian government, greatly expanded in 1855, and operated as the primary meteorological station and official Canadian timekeeper for over fifty years. The Observatory is the country's oldest surviving scientific institution, and is considered the birthplace of Canadian astronomy.

1 May 1969 – Computer chip manufacturer Advanced Micro Devices is started by former executives of Fairchild Semiconductor 3 May 2000 – The outdoor recreation activity Geocaching (logo pictured), in which players use GPS to find hidden containers, is first played 11 May 1976 – The Office of Science and Technology Policy is formed as an Executive Office of the President of the United States 21 May 1844 – After seeing the samples produced by Charles Goodyear, Thomas Hancock becomes the first person to patent vulcanization of rubber. that Tingmissartoq, Charles and Anne Morrow Lindbergh's Lockheed Sirius, was so christened by an Eskimo boy in Greenland, and that its name means "one who flies like a big bird"? ... that an electronic gear-shifting system for bicycles can shift faster than a traditional mechanical system and calibrate itself to minimize maintenance? ... that in 1966, Heinz Waaske created the smallest 135 film camera made to that date, the Rollei--Waleed mufti2 (talk) 04:49, 1 June 2013 (UTC)Information technology (IT) is the application of computers and telecommunications equipment to store, retrieve, transmit and manipulate data,[1] often in the context of a business or other enterprise.[2] The term is commonly used as a synonym for computers and computer networks, but it also encompasses other information distribution technologies such as television and telephones. Several industries are associated with information technology, such as computer hardware, software, electronics, semiconductors, internet, telecom equipment, e-commerce and computer services.[3][4] In a business context, the Information Technology Association of America has defined information technology as "the study, design, development, application, implementation, support or management of computer-based information systems".[5] The responsibilities of those working in the field include network administration, software development and installation, and the planning and management of an organisation's technology life cycle, by which hardware and software is maintained, upgraded, and replaced. Humans have been storing, retrieving, manipulating and communicating information since,in Mesopotamia developed writing in about 3000 BC,[6] but the term "information technology" in its modern sense first appeared in a 1958 article published in the Harvard Business Review; authors Harold J. Leavitt and Thomas L. Whisler commented that "the new technology does not yet have a single established name. We shall call it information technology (IT)."[7] Based on the storage and processing technologies employed, it is possible to distinguish four distinct phases of IT development: pre-mechanical (3000 BC – 1450 AD), mechanical (1450–1840), electromechanical (1840–1940) and electronic (1940–present).[6] This article focuses on the most recent period (electronic), which began in about 1940. Contents.Information technology (IT) is the application of computers and telecommunications equipment to store, retrieve, transmit and manipulate data,[1] often in the context of a business or other enterprise.[2] The term is commonly used as a synonym for computers and computer networks, but it also encompasses other information distribution technologies such as television and telephones. Several industries are associated with information technology, such as computer hardware, software, electronics, semiconductors, internet, telecom equipment, e-commerce and computer services.[3][4] In a business context, the Information Technology Association of America has defined information technology as "the study, design, development, application, implementation, support or management of computer-based information systems".[5] The responsibilities of those working in the field include network administration, software development and installation, and the planning and management of an organisation's technology life cycle, by which hardware and software is maintained, upgraded, and replaced. Humans have been storing, retrieving, manipulating and communicating information since the Sumerians in Mesopotamia developed writing in about 3000 BC,[6] but the term "information technology" in its modern sense first appeared in a 1958 article published in the Harvard Business Review; authors Harold J. Leavitt and Thomas L. Whisler commented that "the new technology does not yet have a single established name. We shall call it information technology (IT)."[7] Based on the storage and processing technologies employed, it is possible to distinguish four distinct phases of IT development: pre-mechanical (3000 BC – 1450 AD), mechanical (1450–1840), electromechanical (1840–1940) and electronic (1940–present).[6] This article focuses on the most recent period (electronic), which began in about 1940. Contents [hide] 1 History of computers 2 Data storage 3 Databases 4 Data retrieval 5 Data transmission 6 Data manipulation 7 Academic perspective 8 Commercial perspective 9 Ethics 10 See also 11 References 12 Further reading 13 External links HISTORY OF COMPUTERS. Devices have been used to aid computation for thousands of years, probably initially in the form of a tally stick.[8] The Antikythera mechanism, dating from about the beginning of the first century BC, is generally considered to be the earliest known mechanical analog computer; it is also the earliest known geared mechanism.[9] Comparable geared devices did not emerge in Europe until the 16th century,[10] and it was not until 1645 that the first mechanical calculator capable of performing the four basic arithmetical operations was developed.[11] Electronic computers, using either relays or valves, began to appear in the early 1940s. The electromechanical Zuse Z3, completed in 1941, was the world's first programmable computer, and by modern standards one of the first machines that could be considered a complete computing machine. Colossus, developed during the Second World War to decrypt German messages was the first electronic digital computer. Although it was programmable, it was not general-purpose, being designed to perform only a single task. It also lacked the ability to store its program in memory. Instead, programming was carried out using plugs and switches to alter the internal wiring.[12] The first recognisably modern electronic digital stored-program computer was the Manchester Small-Scale Experimental Machine (SSEM), which ran its first program on 21 June 1948.[13] DATA STORAGE Main article: Data storage device Early electronic computers such as Colossus made use of punched tape, a long strip of paper on which data was represented by a series of holes, a technology now obsolete.[14] Electronic data storage, which is used in modern computers, dates from the Second World War, when a form of delay line memory was developed to remove the clutter from radar signals, the first practical application of which was the mercury delay line.[15] The first random-access digital storage device was the Williams tube, based on a standard cathode ray tube,[16] but the information stored in it and delay line memory was volatile in that it had to be continuously refreshed, and thus was lost once power was removed. The earliest form of non-volatile computer storage was the magnetic drum, invented in 1932[17] and used in the Ferranti Mark 1, the world's first commercially available general-purpose electronic computer.[18] Most digital data today is still stored magnetically on devices such as hard disk drives, or optically on media such as CD-ROMs.[19] It has been estimated that the worldwide capacity to store information on electronic devices grew from less than 3 exabytes in 1986 to 295 exabytes in 2007,[20] doubling roughly every 3 years.[21] DATA BASES. Database management systems emerged in the 1960s to address the problem of storing and retrieving large amounts of data accurately and quickly. One of the earliest such systems was IBM's Information Management System (IMS),[22] which is still widely deployed more than 40 years later.[23] IMS stores data hierarchically,[22] but in the 1970s Ted Codd proposed an alternative relational storage model based on set theory and predicate logic and the familiar concepts of tables, rows and columns. The first commercially available relational database management system (RDBMS) was available from Oracle in 1980.[24] All database management systems consist of a number of components that together allow the data they store to be accessed simultaneously by many users while maintaining its integrity. A characteristic of all databases is that the structure of the data they contain is defined and stored separately from the data itself, in a database schema.[22] The extensible markup language (XML) has become a popular format for data representation in recent years. Although XML data can be stored in normal file systems, it is commonly held in relational databases to take advantage of their "robust implementation verified by years of both theoretical and practical effort".[25] As an evolution of the Standard Generalized Markup Language (SGML), XML's text-based structure offers the advantage of being both machine and human-readable.[26]. DATA RETRIEVEL The relational database model introduced a programming-language independent Structured Query Language (SQL), based on relational algebra.[24] The terms "data" and "information" are not synonymous. Anything stored is data, but it only becomes information when it is organised and presented meaningfully.[27] Most of the world's digital data is unstructured, and stored in a variety of different physical formats[28][a] even within a single organization. Data warehouses began to be developed in the 1980s to integrate these disparate stores. They typically contain data extracted from various sources, including external sources such as the Internet, organised in such a way as to facilitate decision support systems (DSS).[29] DATA TRANSMISSION Data transmission has three aspects: transmission, propagation, and reception.[30] XML has been increasingly employed as a means of data interchange since the early 2000s,[31] particularly for machine-oriented interactions such as those involved in web-oriented protocols such as SOAP,[26] describing "data-in-transit rather than ... data-at-rest".[31] One of the challenges of such usage is converting data from relational databases into XML Document Object Model (DOM) structures.[32]. DATA MANIPULATION Hilbert and Lopez[20] identify the exponential pace of technological change (a kind of Moore's law): machines' application-specific capacity to compute information per capita roughly doubled every 14 months between 1986 and 2007; the per capita capacity of the world's general-purpose computers doubled every 18 months during the same two decades; the global telecommunication capacity per capita doubled every 34 months; the world's storage capacity per capita required roughly 40 months to double (every 3 years); and per capita broadcast information has doubled every 12.3 years.[20] Massive amounts of data are stored worldwide every day, but unless it can be analysed and presented effectively it essentially resides in what have been called data tombs: "data archives that are seldom visited".[33] To address that issue, the field of data mining – "the process of discovering interesting patterns and knowledge from large amounts of data"[34] – emerged in the late 1980s.[35] ACADEMICS PRESPECTIVE  In an academic context, the Association for Computing Machinery defines IT as "undergraduate degree programs that prepare students to meet the computer technology needs of business, government, healthcare, schools, and other kinds of organizations .... IT specialists assume responsibility for selecting hardware and software products appropriate for an organization, integrating those products with organizational needs and infrastructure, and installing, customizing, and maintaining those applications for the organization’s computer users."[36] COMERCIAL PREPECTIVE The business value of information technology lies in the automation of business processes, provision of information for decision making, connecting businesses with their customers, and the provision of productivity tools to increase efficiency. Worldwide IT spending forecast[37] (billions of U.S. dollars) Category	2012 spending	2013 spending Devices	627	666D Data center systems	141	147 Enterprise software	278	296 IT services	881	927 Telecom services	1,661	1,701 Total	3,588	3,737.