A Brief History of Data

The Story of Data: Past, Present, and (a little) Future

Ankit Rathi
7 min readMay 25, 2024

In this article, I explore the fascinating journey of data throughout human history. From the first marks on bone to the advanced digital systems we use today, the evolution of data is a story of human creativity and progress. By understanding this journey, we can appreciate how important data has been in shaping our world and driving advancements in science, technology, and society.

The first use of data dates back to 19,000 BC when our early human ancestors in the Paleolithic era made use of a tool called the Ishango bone. This tool, made from the bone of a baboon, was used to perform simple calculations. It had notches carved into it, which likely helped in counting or keeping track of numbers. This early form of data recording shows how humans have been using tools to manage information for thousands of years.

In 3400 BC, people in ancient Mesopotamia started using cuneiform writing on clay tablets. They used a reed stylus to make wedge-shaped marks in the clay. This early form of writing was used to record important information like trade transactions, laws, and stories.

Around 3200 BC, the ancient Egyptians began using hieroglyphs, a system of writing that used picture symbols. These hieroglyphs were carved into stone or written on papyrus and used to record important information such as religious texts, official records, and stories.

By 1200 BC, ancient Chinese people used oracle bones to record information. These were usually animal bones or turtle shells that were inscribed with early Chinese characters. They used these bones for divination, asking questions about the future, and recording the answers they believed were revealed.

From the 500s to the 1500s, people in medieval Europe created manuscripts and codices to record information. Monks and scribes painstakingly copied texts by hand onto parchment or vellum, producing books that contained religious texts, scientific knowledge, and literature. These manuscripts were often beautifully decorated with intricate illustrations and served as important tools for preserving and sharing knowledge before the invention of the printing press.

During 1440s, a man named Johannes Gutenberg invented the printing press, a machine that could produce books quickly and efficiently. This invention revolutionized the way information was shared by allowing books to be mass-produced rather than copied by hand. As a result, knowledge became more accessible to a wider audience, leading to significant advancements in education, science, and culture. The printing press is considered one of the most important inventions in history.

Around 1600s, John Graunt looked at a bunch of data and figured out some important stuff. Back then, people didn’t really do much with data, but Graunt changed that. He studied things like birth and death rates in London and noticed patterns. This was a big deal because it helped people understand how diseases spread and how populations grew. Basically, Graunt was one of the first people to use data to learn about the world around us, and his work laid the foundation for modern statistics.

By 1870s, a guy named George Grant made a mechanical calculator. Before this, crunching numbers was a tedious job done by hand. But Grant’s invention changed that. His calculator could add, subtract, multiply, and divide numbers much faster than any human could. It was like the first step toward the calculators we use today. This invention made math a whole lot easier for people, especially scientists and engineers who needed to do lots of calculations. Basically, Grant’s mechanical calculator was a game-changer in making math less of a headache for everyone.

Around 1890s, Herman Hollerith came up with a tabulating machine. This machine was a big deal because it could process a ton of data superfast. How did it work? Well, it used punch cards to store information, kind of like the ones used in early computers. Each hole in the card represented a different piece of data. When you ran the cards through the machine, it could tally up all the info in no time. This invention revolutionized things like census-taking and data processing, making life a whole lot easier for government agencies and businesses.

And 1900s onwards, Fritz Pfleumer introduced magnetic tape. This invention changed the game for storing and accessing data. Basically, magnetic tape is a long strip of plastic coated with a magnetic material. You could record information onto it by magnetizing tiny sections of the tape. What’s cool was that you could store a lot of data on a single reel of magnetic tape, making it really efficient. This technology became widely used for things like audio recording, data storage, and even early computer systems. It paved the way for more advanced forms of data storage we use today, like hard drives and flash memory.

By 1940s, something groundbreaking happened: electronic computers were born. These were machines that could perform complex calculations using electrical circuits instead of mechanical parts. Before this, computers were huge and mainly used for specialized tasks, like decoding messages during wartime. But with electronic computers, things changed fast. They became smaller, faster, and more versatile. Suddenly, scientists, engineers, and businesses had a powerful tool for solving all kinds of problems, from crunching numbers to processing data. The birth of electronic computers marked the beginning of the digital age, shaping the world we live in today.

In the 1970s, E. F. Codd came up with relational databases. These were systems for organizing and managing large amounts of data in a way that made it easy to access and analyze. Before relational databases, storing and retrieving data was a bit of a mess. But Codd’s idea changed that. He proposed a system where data was stored in tables, with each table containing related information. This made it much simpler to query the data and extract the information you needed. Relational databases quickly became the standard for businesses and organizations, laying the foundation for modern data management systems and powering everything from banking systems to social media platforms (after the invention of internet).

During 1990s, Tim Berners-Lee introduced something that transformed the world: the World Wide Web. Before this, sharing information on the internet was pretty clunky and limited to technical users. But Berners-Lee’s invention changed all that. He created a system where documents and resources could be linked together through hypertext, making it easy to navigate from one page to another. This opened up the internet to everyone, not just computer experts. Suddenly, people could access information, communicate, and collaborate on a global scale like never before. The World Wide Web became the backbone of the modern internet, shaping how we learn, work, and connect with each other every day.

At the starting of 2000s, two major technological trends emerged that revolutionized how we handle data: Big Data and Cloud Computing. Big Data refers to the massive amounts of information generated from various sources like social media, sensors, and online transactions. With advancements in technology, we gained the ability to collect, store, and analyze this data to uncover valuable insights and patterns. Cloud Computing, on the other hand, revolutionized how we store and process data by allowing us to access computing resources and services over the internet. Instead of relying solely on local servers, businesses and individuals could now utilize the vast computing power and storage capabilities of remote data centers. Together, these innovations transformed industries, enabling organizations to make data-driven decisions more efficiently and effectively while also democratizing access to powerful computing resources.

Throughout the 2010s, two game-changing technologies surged into the spotlight: Machine Learning and Artificial Intelligence (AI). Machine Learning involves creating algorithms that can learn from data and improve over time without being explicitly programmed. This enabled computers to perform tasks like recognizing patterns, making predictions, and solving complex problems with remarkable accuracy. AI, on the other hand, refers to the broader concept of creating machines or systems that can mimic human intelligence. From virtual assistants like Siri and Alexa to self-driving cars and recommendation systems, AI became deeply integrated into our daily lives, enhancing convenience, efficiency, and decision-making. These advancements sparked a wave of innovation across industries, from healthcare and finance to transportation and entertainment, reshaping how we interact with technology and opening up new frontiers of possibility for the future.

Since the 2020s, two key technologies are taking center stage: Blockchain and Decentralized Data. Blockchain is like a digital ledger that stores records of transactions across a network of computers. Each transaction is encrypted and linked to the previous one, creating a chain of blocks that is secure and transparent. This technology gained popularity for its use in cryptocurrencies like Bitcoin, but its potential extends far beyond that. Blockchain has the power to revolutionize industries by providing secure and tamper-proof systems for tracking assets, verifying identities, and facilitating peer-to-peer transactions without the need for intermediaries. Decentralized Data goes hand in hand with Blockchain, allowing information to be stored and shared across a network of computers rather than relying on a single central authority. This promotes transparency, security, and data sovereignty, empowering individuals and organizations to have greater control over their digital assets and information. Together, Blockchain and Decentralized Data are poised to disrupt traditional systems and pave the way for a more decentralized, transparent, and secure digital future.

From ancient marks to modern technology, we’ve seen how humans have used data to understand the world and make progress. Each step, like John Graunt’s analysis in the 1600s or Tim Berners-Lee’s World Wide Web in the 1990s, has brought new ways to use information. Now, with Blockchain and Decentralized Data in the 2020s, we’re entering a new phase. These technologies promise more security and control over our data. The history of data is still unfolding, showing how we’ve always used information to move forward.

If you loved this story, please feel free to check my other articles on this topic here: https://ankit-rathi.github.io/data-ai-concepts/

Ankit Rathi is a data techie and weekend tradevestor. His interest lies primarily in building end-to-end data applications/products and making money in stock market using Tradevesting methodology.

--

--

Ankit Rathi
Ankit Rathi

Written by Ankit Rathi

ADHD Parent | Data Techie | Weekend Quantvestor | https://ankit-rathi.github.io

Responses (1)