10 Weird Facts about Internet

1-The First Web Page is Still Online
Yes, a restored copy of the first-ever web page is still online and accessible to the public at its original URL. The very first website ever created, info.cern.ch, is still live today, showcasing CERN’s information about the World Wide Web.
The First Web Page
- URL: The original address is
http://info.cern.ch/hypertext/WWW/TheProject.html. - Location: It was hosted on a NeXT computer at CERN, the European Organization for Nuclear Research, in Switzerland.
- Creator: It was created by Tim Berners-Lee, the inventor of the World Wide Web, and first made publicly available on August 6, 1991.
- Content: The page was purely informational, describing the World Wide Web project itself, how to set up a web server and browser, and how to create web pages using hypertext. It contained no images or complex design, just plain text and hyperlinks.
Preservation
CERN launched a project in 2013 to restore the page to its original state and make it available online as a historical artifact, allowing modern users to experience the very beginning of the web. The physical NeXT computer that served the page is also preserved and on display at CERN.
2-There’s a Website That Reads Your Mind
There are popular websites, such as Akinator (akinator.com) and the original 20Q (20q.net), that appear to “read your mind” by accurately guessing characters, objects, or people you are thinking of.
How They Work
These websites use artificial intelligence (AI) and sophisticated algorithms, not actual mind-reading technology.
- Vast Database: They rely on massive databases of characters and objects, continuously built and updated through millions of user interactions.
- Decision Tree Logic: They employ a decision tree or similar algorithm (like a binary search) to ask a series of yes/no questions. Each answer eliminates a large number of possibilities, quickly narrowing down the options.
- Machine Learning: The systems learn from every game played. If they guess incorrectly, users can provide the right answer, which expands the database and improves the AI’s accuracy for future guesses.
- Probabilistic Reasoning: The questions are chosen dynamically to maximize the information gained at each step, making the guessing process very efficient and often surprisingly accurate, which creates the illusion of mind-reading.
Current Technology
While a website cannot literally access your private thoughts through a standard web connection, AI and neuroscience research is exploring ways to decode brain signals. Recent advancements have demonstrated the ability of AI systems to translate brain activity (recorded via non-invasive EEG caps or brain implants) into text or images, but this requires specialized hardware and is not something a simple website can do.
In summary, the “mind-reading” websites are clever and highly effective guessing games that use large datasets and smart algorithms to create an entertaining illusion of mental prowess.
3-A Single Byte of Data Can Be a Whole Web Page
Yes, a single byte of data can be an entire, valid web page. The smallest possible valid HTML5 page, when stripped to its absolute minimum, can be represented by as few as five characters (five bytes in most encodings), or in highly compressed formats potentially even fewer, depending on how you define “functional”.
Technical Explanation
- Minimal HTML: A modern, valid HTML5 document can be incredibly short. For example, a page that displays a single character can be just that character (e.g., the character “A”). The browser automatically adds the necessary
<html>,<head>, and<body>elements that are technically required by the specification but often optional in practice due to browser tolerance. - The Single Character: A single character, such as “A”, can be stored as one byte using the standard ASCII or UTF-8 character encoding. When a server sends a response containing just this single byte and the correct HTTP headers (which add overhead on the transmission level, but are not part of the page’s data size), the browser will render it as a complete web page with that single character in the top-left corner.
- Data vs. Protocol Overhead: The statement refers to the actual content data of the page itself. The process of requesting and receiving the page involves significant additional overhead from the HTTP and TCP/IP protocols (headers, handshakes, etc.), which are many times larger than a single byte.
- Compression: In theory, you could use compression techniques to represent an extremely large page with complex content using a single byte, but this would only work if the data was highly redundant or the system had pre-defined rules for decompression (e.g., mapping a single byte to a large, pre-stored image or text block). This is a theoretical concept (Kolmogorov complexity) and not how standard web pages function.
4-The Internet We Know Today Was Started by a Military Project
Yes, the Internet as we know it today has its origins in a U.S. military project called ARPANET.
The Military Connection
- Funding and Oversight: The ARPANET (Advanced Research Projects Agency Network) project was established by the U.S. Department of Defense’s Advanced Research Projects Agency (ARPA), now known as DARPA.
- Initial Motivation: The primary initial motivation was not necessarily to survive a nuclear attack, as a popular myth suggests, but to enable resource sharing among remote computers and, later, to create a robust, fault-tolerant communication system that could function even if parts of the network were compromised.
- Key Technologies: The research, which involved collaboration with universities and research institutions, led to the development of fundamental technologies still used today, most notably packet switching and the TCP/IP protocol suite, which forms the core of the modern Internet.
Web Site
- Academic Collaboration: While initiated by the military, the ARPANET quickly expanded to connect research labs and universities to foster collaboration among scientists and academics.
- Separation: In 1983, the military sites were split onto a separate, dedicated Military Network (MILNET), leaving ARPANET for unclassified academic and research purposes.
- NSFNET and Commercialization: The National Science Foundation Network (NSFNET) eventually replaced ARPANET as the primary backbone for university traffic in the mid-1980s. The linking of commercial networks and the invention of the World Wide Web by Tim Berners-Lee at CERN in 1989
- marked the beginning of the Internet’s commercial expansion and public use.

5-There’s a Website for Everything
While not literally true that a website exists for every single conceivable idea, the phrase captures a core truth about the vastness of the internet: there is an extraordinary diversity of websites, covering everything from the mainstream to the incredibly niche and obscure.
The World Wide Web contains over a billion websites, catering to a near-infinite range of human interests, hobbies, and needs.
Range of Topics
- Major Topics: You can find extensive resources on general topics like news, education (Wikipedia, online courses), commerce (Amazon), entertainment (YouTube, social media), and professional networking (LinkedIn).
- Niche Interests: The internet excels at catering to highly specific “niche” topics. Examples of the kind of obscure or specialized websites that exist include:
- Sites dedicated to the care and cultivation of specific types of cacti or unique birdhouse designs.
- Blogs focused purely on the intricacies of retro gaming, adult coloring books, or sustainable fashion.
- Websites offering detailed product reviews for very specific items, such as niche camping gear or specific brands of coffee beans.
- Sites that aggregate long-form journalism, detail curious global landmarks (like Atlas Obscura), or allow you to anonymously post secrets (Post Secret).
- User-Generated Content: Platforms like Reddit and specialized forums host communities and discussions on countless micro-topics that would never have a dedicated, professionally built website, further extending the reach of “everything” online.
6-The ‘404 Error’ Has a Backstory

There is a popular and widely circulated story about the origin of the ‘404 Error’ message, but it is largely considered an urban legend by the people involved in creating the web.
The Mythical Backstory
The most common story suggests that the original, central database for the World Wide Web was housed in an office on the fourth floor of a building at CERN (the European Organization for Nuclear Research) in Switzerland, specifically in Room 404.
According to this legend, in the very early days of the web, two or three people were tasked with manually locating and transferring requested files. When they couldn’t find a requested file (often due to a user typo), they would send back a message: “Room 404: file not found”. When the process was automated, the number 404 stuck as the standard error code.
The Actual Origin
Robert Cailliau, a key figure in the development of the World Wide Web alongside Tim Berners-Lee, has explicitly debunked the Room 404 myth. He stated that:
- The number 404 was never linked to any physical room or place at CERN; there wasn’t even a Room 404 in the relevant building at the time.
- The system of HTTP status codes was created as a straightforward, efficient way to categorize errors within the web protocol.
- The first ‘4’ designates a client error (meaning the user’s request was faulty, like a bad URL), as opposed to a server error (5xx codes) or a successful request (2xx codes).
- The subsequent numbers were assigned somewhat arbitrarily by the programmers within the 40x range to signify the specific error of “Not Found”.
7-The Internet Has Its Own Ecosystem
Yes, the term “Internet ecosystem” is widely used by organizations like the Internet Society to describe the complex, interconnected web of organizations, technologies, and communities that all contribute to the internet’s function, evolution, and governance.
Components of the Internet Ecosystem
Like a natural ecosystem, the internet is not owned or controlled by a single entity, but rather relies on the collaboration and interaction of diverse participants and components.
Key components include:
- Physical Infrastructure: This is the abiotic (non-living) part, comprising the physical hardware that makes data move, such as fiber optic cables (including vast undersea cables that connect continents), routers, switches, cell towers, satellites, and massive data centers and servers.
- Protocols and Standards: These are the rules (like TCP/IP, HTTP, DNS) that allow different technologies and networks to communicate with each other. These open standards are developed and maintained by non-profit organizations like the Internet Engineering Task Force (IETF).
- Organizations and Communities: This represents the biotic (living) element. It includes Internet Service Providers (ISPs), equipment manufacturers, content providers (websites, apps), research institutions, governments, and a global community of engineers, users, and policy-makers who help the internet work and evolve.
- Users: The billions of people and devices at the “edges” of the network that generate requests, consume data, and drive the need for the system to constantly adapt and grow.
8-Most of the Internet’s Data Is Old
That statement is incorrect; in reality, the vast majority of the internet’s data is relatively new, with massive amounts of new information generated every single day.
The Growth of Data
The volume of data created and consumed globally is expanding exponentially.
- 90% in Two Years: It is estimated that around 90% of the world’s data was generated in just the last two years alone. This demonstrates the incredible speed at which new data is produced compared to the total existing archive.
- Daily Generation: Hundreds of millions of terabytes of new data are generated every single day through emails, social media interactions, online transactions, streaming media, and machine-to-machine communication.
- Dynamic Content: Much of the web is dynamic (constantly changing), driven by social media feeds, live updates, and e-commerce, which means data has a short shelf life before being updated or archived.
9-There’s a Website That Simulates the Entire Internet
No, there is no single website that simulates the entire internet. The internet is far too vast, complex, and dynamic for a complete, accurate simulation in one place.
However, several websites and tools exist that provide different types of “internet simulation” for specific purposes:
Types of Internet Simulation Websites/Tools
- AI-Powered Web Simulators (e.g., WebSim): Tools like WebSim AI allow users to generate functional, interactive websites or applications within a simulated, AI-created “alternative internet” using simple text prompts or URLs. These generated sites only exist within that specific environment and do not mirror real internet content, but demonstrate AI’s ability to mimic web pages.
- Historical Browsing (e.g., oldweb.today): Websites like oldweb.today act as time machines. They use emulated versions of old web browsers (like Netscape Navigator or early Internet Explorer) to access archived snapshots of old websites from the Internet Archive, providing a simulation of what browsing was like in the past.
- Network Training Simulators: For educational or professional use, there are network simulators (like Cisco Packet Tracer or Code.org’s Internet Simulator) that allow users to learn how networks function, experiment with protocols, and understand how data moves, but they do not contain the content of the entire internet.
- Testing and Development Tools: Developers use various tools and services to simulate how their websites will appear and function on different devices, operating systems, and browsers (e.g., Browserling, BrowserStack). These are for testing specific websites, not the internet as a whole.
10-The Deep Web is Much Larger Than the Surface Web

Yes, the statement is correct: the Deep Web is vastly larger than the Surface Web (the part of the internet indexed by search engines like Google).
Key Comparisons
- Surface Web: This is the “visible” part of the internet that you access daily via search engines. It’s estimated to make up only about 4% to 10% of the entire web.
- Deep Web: This comprises everything not indexed by search engines, making up an estimated 90% to 96% of the internet. Some estimates suggest the deep web is 400 to 550 times larger than the surface web.
Deep Web vs. Dark Web
It is important not to confuse the Deep Web with the Dark Web.
which is a much smaller, highly encrypted subset of the Deep Web requiring specialized software like the Tor browser to access. The Dark Web is often associated with illicit activities, whereas the majority of the Deep Web is not.
Thanks for sharing. I read many of your blog posts, cool, your blog is very good. https://accounts.binance.info/vi/register-person?ref=MFN0EVO1
Thnks