The History of Computers

While computers are now an important part of the lives of human beings, there was a time where computers did not exist. Knowing the history of computers and how much progression has been made can help you understand just how complicated and innovative the creation of computers really is.

Unlike most devices, the computer is one of the few inventions that does not have one specific inventor. Throughout the development of the computer, many people have added their creations to the list required to make a computer work. Some of the inventions have been different types of computers, and some of them were parts required to allow computers to be developed further.

The Beginning

Perhaps the most significant date in the history of computers is the year 1936. It was in this year that the first “computer” was developed. It was created by Konrad Zuse and dubbed the Z1 Computer. This computer stands as the first as it was the first system to be fully programmable. There were devices prior to this, but none had the computing power that sets it apart from other electronics.

It wasn’t until 1942 that any business saw profit and opportunity in computers. This first company was called ABC computers, owned and operated by John Atanasoff and Clifford Berry. Two years later, the Harvard Mark I computer was developed, furthering the science of computing.

Over the course of the next few years, inventors all over the world began to search more into the study of computers, and how to improve upon them. Those next ten years say the introduction of the transistor, which would become a vital part of the inner workings of the computer, the ENIAC 1 computer, as well as many other types of systems. The ENIAC 1 is perhaps one of the most interesting, as it required 20,000 vacuum tubes to operate. It was a massive machine, and started the revolution to build smaller and faster computers.

The age of computers was forever altered by the introduction of International Business Machines, or IBM, into the computing industry in 1953. This company, over the course of computer history, has been a major player in the development of new systems and servers for public and private use. This introduction brought about the first real signs of competition within computing history, which helped to spur faster and better development of computers. Their first contribution was the IBM 701 EDPM Computer.

A Programming Language Evolves

A year later, the first successful high level programming language was created. This was a programming language not written in ‘assembly’ or binary, which are considered very low level languages. FORTRAN was written so that more people could begin to program computers easily.

The year 1955, the Bank of America, coupled with Stanford Research Institute and General Electric, saw the creation of the first computers for use in banks. The MICR, or Magnetic Ink Character Recognition, coupled with the actual computer, the ERMA, was a breakthrough for the banking industry. It wasn’t until 1959 that the pair of systems were put into use in actual banks.

During 1958, one of the most important breakthroughs in computer history occurred, the creation of the integrated circuit. This device, also known as the chip, is one of the base requirements for modern computer systems. On every motherboard and card within a computer system, are many chips that contain information on what the boards and cards do. Without these chips, the systems as we know them today cannot function.