Fact-checked by Grok 2 weeks ago

Text

Text is the original words and form of a written or printed work, presented as a main body of content distinct from supplementary elements like illustrations or annotations. It encompasses any coherent sequence of characters, words, or sentences that convey meaning, serving as the primary medium for human communication, documentation, and expression across cultures and eras. In its broadest sense, text includes both verbal records of spoken discourse and fixed written compositions, forming the foundation of literature, legal documents, scientific papers, and everyday messaging. In , text is understood as a semantic unit of in use—a passage, spoken or written, of any length that constitutes a unified whole through mechanisms like and context. This unity arises from semantic relations that bind elements together, distinguishing a text from mere disjointed ; for instance, cohesive devices such as , , and lexical create , ensuring interpretability as a single communicative event. , a subfield emerging in the mid-20th century, analyzes these structures to explore how texts function beyond the level, classifying them by types like , argumentative, or descriptive based on purpose and organization. Key principles include (logical sense-making) and (producer's aim to communicate), which apply to both oral traditions and modern media. With the advent of digital technologies, text has expanded into computational realms, where it denotes a of encoded characters—typically in formats like ASCII or —processable by machines for storage, search, and display. This digital text underpins applications from word processors to , enabling global while raising issues like , encoding standards, and algorithmic analysis in fields such as . Overall, text remains a cornerstone of preservation and , adapting from ancient scripts on clay tablets to contemporary hyperlinked documents.

Fundamentals

Definition and Scope

Text is fundamentally a sequence of symbols or characters, including letters, numbers, , and other graphical elements, organized to represent and convey meaning or . This arrangement forms a coherent that can be interpreted within a specific . While purely ephemeral spoken relies on auditory delivery without fixed persistence, text often provides visual or recorded fixity, encompassing verbal records in . Unlike images or visual , which communicate through spatial and pictorial representation, text operates as a linear or structured progression of linguistic signs, enabling precise encoding of ideas, instructions, or narratives. In , text extends to coherent spoken passages or their records, forming a semantic bound by and . In , the study of signs and their interpretation, text exemplifies a system of signifiers—the material forms of characters or symbols—and signifieds—the concepts or meanings they evoke, creating a relational structure that generates communicative intent. This semiotic framework underscores text's role as a mediated artifact, where meaning emerges not from isolated symbols but from their contextual interplay, as articulated in foundational theories of signification. Text thus serves as a primary vehicle for human expression, bridging individual with shared cultural understanding. The scope of text extends across diverse manifestations, encompassing handwritten manuscripts, printed books, digital documents, computer code, and public signage, all of which facilitate the storage, transmission, and dissemination of . It plays a pivotal role in by enabling thought to be externalized and preserved across generations, from ancient inscriptions on clay tablets to contemporary hypertext systems in online environments that link interconnected nodes of information. This breadth highlights text's enduring function in cultural preservation, allowing societies to document history, laws, and innovations for future access and analysis. Text is delineated from non-textual data, such as audio recordings or video streams, which transmit via temporal auditory or visual sequences without relying on discrete symbolic alphabets or grammars inherent to linguistic text. This distinction emphasizes text's static, decodable nature, optimized for rereading and precise replication, in contrast to the dynamic, sensory immersion of forms.

Etymology and Evolution

The word "text" originates from the Latin textus, meaning "woven fabric" or "structure," derived from the verb texere, "to weave," which evokes the idea of words interlaced like threads in a fabric. This root emphasized the interconnectedness of written content, initially applied in medieval Latin to denote the Scriptures or a treatise as a cohesive, woven whole. By the 12th century, it entered Old French as texte, referring to written passages, and reached Middle English around 1300 as text, primarily signifying the words of an author or sacred writings, such as biblical verses. In this early usage, text carried a sense of narrative structure, mirroring the weaving metaphor in its portrayal of stories or doctrines as intricately bound elements. Conceptually, the notion of text evolved from the sacred, handcrafted weave of medieval manuscripts, where scribes produced unique codices that intertwined illumination, commentary, and scripture into a singular artifact, to the reproducible content enabled by 19th-century advancements. In , text was viewed as a holy or authoritative fabric, often glossed with marginal notes that "wove" additional layers of interpretation, reinforcing its role as a dynamic, interpretive structure rather than mere words. The invention of Johannes Gutenberg's movable-type in the 1450s marked a pivotal milestone, transforming text into a mass medium by allowing rapid, identical reproduction of , which democratized access to and shifted emphasis from artisanal uniqueness to standardized dissemination. This evolution continued into the , expanding text beyond linear forms; in 1965, coined "hypertext" to describe non-sequential, linked writing systems, influencing digital navigation and challenging the traditional woven linearity of print. Post-2000 digital developments further extended the boundaries of text, incorporating visual and interactive elements like , which originated in mobile messaging in the late 1990s but proliferated globally after standardization in 2010, functioning as symbolic extensions that enrich textual communication with emotional and contextual nuance. Philosophically, Julia Kristeva's theory of , introduced in the late through essays like "Word, Dialogue and ," reframed text as an interwoven mosaic of references to prior works, drawing on Mikhail Bakhtin's dialogism to argue that no text exists in isolation but builds upon and transforms a cultural "fabric" of allusions. This perspective underscores the ongoing evolution of text as a relational, absorptive entity, adapting from sanctity to digital interconnectivity.

Written Text

History of Written Text

The origins of written text trace back to prehistoric symbolic representations, with cave paintings and pictographs emerging around 40,000 BCE serving as early forms of proto-text that conveyed meaning through visual symbols, though not yet systematic writing. These evolved through administrative tools like clay tokens used in Mesopotamia from 8000 to 3500 BCE for counting goods, marking a precursor to true writing systems. By approximately 3200 BCE, the Sumerians in Mesopotamia developed cuneiform, the earliest known full writing system, initially as pictographic impressions on clay tablets for accounting purposes such as recording grain, livestock, and labor in urban centers like Uruk. This wedge-shaped script, created with a reed stylus, began with around 1,200 signs and facilitated the bureaucracy of early city-states, representing a shift from symbolic notation to linguistic recording. Major writing systems soon diversified across regions. In , hieroglyphs appeared around 3100 BCE as a logographic and phonetic script carved on stone or written on , used for religious, ceremonial, and administrative texts in tombs and monuments. Independently invented from , this system combined pictorial ideograms with sound signs, enduring for over 3,000 years. In , logographic writing emerged by 1200 BCE during the , with inscribed on animal bones and turtle shells for , featuring thousands of characters representing words or morphemes that form the basis of modern Chinese hanzi. Non-Western systems like the glyphs, a logosyllabic script developed in by around 300 BCE (with precursors possibly earlier), , astronomy, and rituals on stone, bark paper, and , though it remained undeciphered until breakthroughs in the 1950s by scholars like , who identified phonetic elements using the . The , originating around 1050 BCE in the , introduced a revolutionary consonantal system of 22 letters, which influenced the Greek alphabet (adding vowels by 800 BCE) and, through it, the that underpins many modern Western languages. Key technological inventions advanced the form and dissemination of written text. Papyrus scrolls, invented in ancient Egypt around 3000 BCE from the plant, provided a lightweight, flexible surface superior to clay for longer texts, enabling the production of rolls up to 40 meters long for literature and records. By the 1st century CE, the codex format—pages bound along one edge, often from or —began replacing scrolls in the world, offering easier access to content via random page turning and greater durability, with early literary examples appearing among Christian communities. Meanwhile, in around 105 CE, court official invented by processing mulberry bark, hemp fibers, rags, and fishnets into thin sheets, offering a cheaper and more portable medium than bamboo slips or silk and greatly expanding access to written materials. In , Bi Sheng invented movable type printing around 1040 CE using fired clay characters that could be individually arranged in a frame and reused to compose different texts, though it saw limited adoption compared to woodblock methods until the 15th century. Cultural and trade networks propelled the spread of writing systems. Along the from the 2nd century BCE, scripts and printing techniques diffused eastward and westward, with reaching Central Asian Uighurs by the 11th century and influencing regional adaptations like Sogdian and Uighur systems derived from forms. Colonial expansions, particularly European from the 15th to 19th centuries, imposed Latin-based alphabets on indigenous systems worldwide, standardizing scripts in the , , and through missionary and administrative efforts, often supplanting local traditions like Mayan glyphs.

Typography and Layout

Typography encompasses the visual arrangement of text to enhance , convey meaning, and achieve aesthetic appeal in print and contexts. Central to this are core principles governing typefaces, spacing, and positioning. Typefaces are categorized broadly into and varieties: fonts feature small decorative strokes at the ends of letterforms, aiding eye flow in printed materials like books, while fonts offer a cleaner, more modern look suited to and digital interfaces. Key spacing techniques include , which adjusts the space between specific pairs of letters to prevent awkward gaps, and leading, the vertical space between lines that ensures comfortable reading without crowding. Alignment further refines , with left alignment promoting natural reading rhythms in Western languages, justified alignment creating uniform edges for formal documents, and centered or right alignment used sparingly for emphasis. These elements originated in historical scripts and evolved through innovations. The historical development of typefaces traces from the 12th-century , or , a dense, angular style mimicking medieval manuscripts and popularized in early printed works like Gutenberg's Bible. This gave way to more legible forms in the , culminating in the late 18th century with Giambattista Bodoni's contributions; as an Italian printer and type designer (1740–1813), Bodoni pioneered high-contrast, neoclassical typefaces emphasizing sharp thin-thick stroke variations and geometric precision, influencing modern designs. By the mid-20th century, innovations like —designed in 1957 by for the Haas Type Foundry—emerged as a neutral, versatile standard for corporate and editorial use, reflecting post-war modernist ideals. Layout techniques focus on structuring text for clarity and flow, including justification, which evenly spaces words across a line to create straight edges on both sides, margins that frame content with to reduce visual fatigue, and , where larger headings and varied weights distinguish sections from body text. underpin these, dividing pages into modular units for consistent alignment; Jan Tschichold's 1928 manifesto Die neue Typographie advocated asymmetric grids, type, and functional layouts, revolutionizing by prioritizing content over ornamentation. The evolution of typographic tools progressed from hand-lettering, where scribes manually formed letters using tools like quills on , to mechanical methods in the . Phototypesetting in the introduced photographic exposure of text onto film, enabling faster composition than metal type. The digital era began with Adobe's in 1982, which standardized vector-based font rendering, allowing scalable, high-quality type across devices and paving the way for . Typographic standards ensure consistency and precision, such as the use of em dashes (—) for interruptions or parenthetical phrases and en dashes (–) for ranges or connections, as outlined in authoritative style guides. Accessibility considerations, particularly post-2000 guidelines from organizations like the , recommend fonts in 12–14 point sizes, increased leading, and papers to mitigate visual stress for dyslexic readers. Recent advancements address efficiency and : variable fonts, formalized in the 2016 OpenType 1.8 specification and integrated into W3C CSS standards, enable a single file to vary weight, width, and other axes dynamically, reducing file sizes and improving . In the 2020s, eco-trends emphasize sustainable inks derived from soy or , alongside "eco-fonts" designed with thinner strokes to cut ink usage by up to 28% in print runs, aligning typography with environmental goals.

Linguistic Aspects

Text, as a written representation of language, encompasses structural elements that organize linguistic units into coherent forms. Syntax governs the arrangement of words and phrases to form sentences, ensuring grammaticality and conveying relational meanings, such as subject-verb-object order in English. Morphology deals with the internal structure of words, including processes like inflection (e.g., adding -s for plurals) and derivation (e.g., un- prefixing to form negatives), which modify meaning and grammatical function. Orthography, the conventional spelling system, approximates phonetics by mapping written symbols to spoken sounds, though inconsistencies arise, as in English where "ough" can represent diverse pronunciations across words like "through," "cough," and "though." These elements interact to produce readable text, bridging spoken language with its written counterpart. Semantics and pragmatics further shape how text conveys meaning beyond literal structures. Semantics focuses on the denotative meanings of words and sentences, addressing phenomena like , where a single form holds multiple related senses—such as "" referring to a or a —resolved through contextual clues. Pragmatics examines implied meanings derived from situational context, speaker intent, and cultural norms, enabling inferences like sarcasm in written . Discourse analysis extends this to larger textual units, such as paragraphs, where cohesion through pronouns, conjunctions, and thematic progression creates unified narratives or arguments, influencing interpretation in genres from essays to narratives. Linguistic variations in text reflect sociolinguistic diversity and evolution. Dialectal differences manifest in spelling and vocabulary, as seen in British English ("colour," "lift") versus American English ("color," "elevator"), stemming from historical divergences post-colonization and standardized through publishing norms. Code-switching occurs in multilingual texts, where writers alternate languages within sentences or paragraphs to convey identity, emphasize points, or accommodate audiences, common in bilingual communities and literature. Textual evolution incorporates slang, particularly internet acronyms emerging post-1990s like "LOL" (laughing out loud) and "BRB" (be right back), driven by digital communication's brevity needs and rapidly integrating into formal writing. Theoretical frameworks underpin these aspects, notably Ferdinand de Saussure's structuralism, which distinguishes langue (the abstract system of language rules) from parole (individual acts of usage), applying to written text as a fixed manifestation of langue that enables shared interpretation. Noam Chomsky's generative grammar posits innate universal principles generating syntactic structures, facilitating text parsing by modeling hierarchical phrase structures and transformations, influencing how readers intuitively process sentence complexity. In computational linguistics, these intersect with natural language processing (NLP) basics, where traditional linguistic models inform text analysis tasks like parsing without delving into algorithms, enhancing machine understanding of human language patterns. The rise of in text generation introduces ethical linguistic concerns, such as biases in generated content perpetuating dialectal or diluting cultural nuances in multilingual outputs, prompting calls for inclusive training data to preserve sociolinguistic . These tools challenge traditional authorship by mimicking human-like semantics and , raising issues of originality and equitable representation in written .

Digital Text

Representation in Computing

In computing, text representation began with the American Standard Code for Information Interchange (ASCII), a 7-bit encoding standard published in 1963 by the (ANSI) to support 128 characters primarily for the , including uppercase and lowercase letters, digits, , and control codes. This limited scope addressed early digital needs but excluded non-Latin scripts. To extend ASCII for Western European languages, the International Organization for Standardization (ISO) developed the ISO/IEC 8859 family of 8-bit encodings in the 1980s, each variant adding 128 characters for specific linguistic needs, such as accented letters in ISO 8859-1 (Latin-1). The standard, initiated by the in 1991, revolutionized text representation by providing a universal encoding for over 159,000 characters across all writing systems as of version 17.0 in 2025. Version 17.0 further expanded this with 4,803 new characters, including four new scripts and additional emoji. Among its transformation formats, emerged as the dominant encoding due to its with ASCII—using a single byte for the first 128 characters—and variable-length encoding (1 to 4 bytes per character) for others, enabling efficient storage of multilingual text. By 2023, accounted for approximately 98.8% of websites with known character encodings, underscoring its prevalence in and software development. In data structures, text is typically stored as , which are contiguous arrays of where each element represents a in the chosen encoding. In , string length in bytes differs from character count due to variable byte usage—for instance, English text uses 1 byte per character, while may require 3 bytes—necessitating careful handling in memory allocation and indexing. To ensure consistent representation and avoid equivalence issues (e.g., precomposed vs. decomposed accented characters), defines normalization forms: Normalization Form C (NFC) composes compatible characters into single code points for compact storage, while Normalization Form D (NFD) decomposes them for search and sorting applications. Text rendering involves converting encoded characters into visual glyphs, starting with font rasterization, where vector-based font outlines (e.g., in or formats) are sampled and filled into pixel grids to match display resolutions, often using for smoothness. For languages with right-to-left (RTL) scripts like , the Unicode Bidirectional Algorithm (UBA), specified in Standard Annex #9 and first detailed in 1999, determines text directionality by resolving embedding levels and reordering mixed left-to-right (LTR) and RTL segments to produce correct visual order. The evolution of text representation traces from punched cards in the 1890s, where Herman Hollerith's system encoded data via holes in 80-column cards for the U.S. Census, representing characters through mechanical tabulation, to the 1990s introduction of , which leveraged ASCII and early for structured web text display. Challenges persisted, such as integrating non-alphabetic symbols; 6.0 in 2010 introduced formal support for over 700 characters, enabling their global use in digital communication. Emerging research in 2025 explores quantum computing's implications for text representation, where quantum bits (qubits) could enable superposition-based encodings for probabilistic text models in , potentially surpassing classical limits in handling ambiguous or vast character sets, though practical implementations remain experimental.

Transmission and Telecommunications

The of text has evolved from mechanical and electrical systems to sophisticated digital networks, enabling rapid global communication. Early methods relied on coded electrical signals to convey textual information over wires. In 1837, and developed the telegraph system using , a series of dots and dashes representing letters and numbers, which allowed for the first long-distance of text-based messages. By the , teletype machines automated this , using perforated and electric impulses to send and receive text at speeds up to 60 words per minute over telephone lines, forming the basis for early automated messaging services. technology, initially invented by Alexander Bain in 1843 as a mechanical scanning device for transmitting images including text, saw digital advancements in the 1960s with the introduction of standards like Group 3, enabling electronic faxing over phone lines at resolutions of 200 dpi. Modern text transmission is governed by standardized protocols that facilitate efficient data exchange over the and mobile networks. The (SMTP), defined in 821 in 1982, established the foundation for by specifying how text messages are routed between servers using TCP/IP. For web-based text, the Hypertext Transfer Protocol (HTTP), proposed by in 1991 and formalized in 1945, enables the retrieval and display of textual content in hyperlinked documents. (SMS), introduced in 1992 on the network, standardized with a 160-character limit to fit within 140 bytes of data, supporting alphanumeric content and emoticons. protocols like Extensible Messaging and Presence Protocol (XMPP), developed in 1999 as an by the Software Foundation, allow real-time text exchange in chat applications through XML-based streams. To optimize and protect data, transmission incorporates and measures tailored to text's structured nature. , introduced in 1992 and widely adopted for HTTP in the 1990s per 1952, compresses textual content using the algorithm, achieving ratios of 2:1 to 10:1 for and similar formats by exploiting redundancy in strings. For , (TLS), succeeding SSL and specified in 2246 in 1999, encrypts text in transit using asymmetric and symmetric , ensuring confidentiality and integrity in protocols like and secure email. These advancements have profound global impacts, particularly through and over-the-top () services. Unicode's adoption in post-2000, via standards like in mobile protocols, enabled seamless transmission of multilingual text, supporting over 149,000 characters across scripts. applications, such as launched in 2009, have revolutionized by bypassing traditional carriers; by 2025, it handles over 100 billion messages daily, leveraging networks for end-to-end encrypted chats. In the 2020s, networks reduced text transmission to under 1 for low-latency applications, enhancing messaging in and mobile scenarios. Additionally, satellite integration like SpaceX's in 2024 extended to remote areas, providing global coverage with low-earth-orbit relays for resilient communication.

Processing and Analysis

Processing and analysis of digital text encompass a range of computational techniques designed to manipulate, search, and interpret textual data after it has been encoded. Basic operations form the foundation of these processes. Searching within text often relies on regular expressions (regex), which allow for pattern matching using formalized notations originally developed by Stephen Kleene in 1956 to describe regular events in automata theory. These were adapted for practical computing by Ken Thompson in 1968 for the Unix editor 'ed', enabling efficient identification of substrings based on complex criteria like character classes or repetitions. Tokenization, the process of splitting text into smaller units such as words or sentences, is another essential operation, facilitating subsequent analysis by breaking down raw input into manageable tokens while handling punctuation and whitespace. Sorting text, typically alphabetically or by relevance, supports tasks like indexing and retrieval, often integrated into database queries or file management systems. Advanced techniques in (NLP) build upon these basics to extract meaning and insights from text. , which determines the emotional tone of text (positive, negative, or neutral), gained prominence in the 2010s with the rise of , evolving from lexicon-based methods to models trained on large datasets. , exemplified by launched in 2006, uses statistical and neural approaches to convert text between languages, initially relying on parallel corpora from sources like documents. These methods enable applications such as automated summarization and , where models infer context and intent from unstructured text. Key tools and libraries support these operations at scale. Regular expressions are implemented in languages like and , providing built-in support for pattern-based searching since the 1970s in Unix utilities like . The Natural Language Toolkit (NLTK), released in 2001, offers Python-based APIs for tokenization, , and basic NLP tasks, serving as an educational and staple with access to corpora like Penn Treebank. For , , introduced in 2006, enables distributed processing of massive text corpora across clusters, using to handle tasks like indexing petabyte-scale document collections. Large language models (LLMs) represent a high-impact advancement in text generation and analysis, capable of producing coherent, context-aware text from prompts. The GPT series, starting with in 2018 from , pioneered generative pre-training on vast internet-scale data, achieving breakthroughs in tasks like story completion and through architectures. Subsequent models like (2020) scaled to billions of parameters, enabling where models perform unseen tasks via natural language instructions. By 2025, ethical updates in LLM development emphasize responsible deployment, including benchmarks for to address societal harms. Challenges in text processing persist, particularly in resolving ambiguities and mitigating biases. Ambiguity resolution, such as , requires contextual analysis to distinguish meanings (e.g., "" as versus river edge), often using supervised models trained on sense-annotated corpora like SemCor. Bias in AI text models, amplified in LLMs trained on internet data, can perpetuate stereotypes in generated outputs, prompting fairness audits in the 2020s that evaluate demographic parity across genders, races, and cultures. These audits, involving techniques like counterfactual testing, aim to quantify and reduce disparities, ensuring equitable performance in real-world applications like hiring tools or .

Text in Arts and Culture

Literature and Narrative

Literature and narrative represent text as a primary medium for , encompassing diverse forms that have evolved from ancient oral traditions to contemporary digital expressions. The transition from oral epics to written marked a pivotal shift, enabling the preservation and dissemination of narratives across cultures. One of the earliest examples is the , an Mesopotamian epic poem composed in written form around 2100 BCE, which explores themes of heroism, friendship, and mortality through tablets. This oral-to-written evolution laid the foundation for structured literary genres, influencing subsequent narrative traditions worldwide. Key forms of literary text include , , and , each employing text to convey complex narratives and emotions. The emerged as a prominent prose form with Miguel de Cervantes's , published in 1605, widely regarded as the first modern for its innovative use of irony, , and character depth, blending with . , often concise and rhythmic, captures introspective or lyrical narratives, while structures text for performance, as seen in tragedies or Elizabethan plays. These genres have adapted over time, with printed like democratizing access to long-form storytelling following the invention of the . Narrative techniques in literature utilize text to manipulate perspective, pacing, and structure, enhancing reader engagement. Point of view—ranging from first-person intimacy to omniscient detachment—shapes how events unfold, allowing authors to reveal or withhold information. Plot structure often follows models like Gustav Freytag's pyramid, outlined in his 1863 work Die Technik des Dramas, which divides narratives into exposition, rising action, climax, falling action, and denouement to create dramatic tension. A groundbreaking technique is stream-of-consciousness, pioneered in James Joyce's Ulysses (1922), which mimics the fluid, associative flow of human thought through unpunctuated, fragmented prose, immersing readers in characters' inner worlds. Literature's cultural roles extend to canon formation, where texts are selected and preserved as exemplars of artistic value, often reflecting societal biases. , dominated by European works like , has historically marginalized global traditions, yet collections such as —a frame narrative compiled in Arabic during the —highlight diverse storytelling from Middle Eastern, , and sources, influencing universal motifs like Scheherazade's tale-telling for survival. Translations play a crucial role in dissemination, bridging linguistic barriers to globalize literature; for instance, data from 2000–2009 shows translations facilitate cultural exchange, with non-Western works gaining visibility through rendered versions that preserve narrative essence while adapting idioms. This process challenges Eurocentric canons, promoting a more inclusive global literary heritage. Digital literature expands possibilities through interactive and user-generated text, emerging in the with on personal computers, such as Infocom's parser-based adventure games that allowed readers to influence plot via commands. Fanfiction, derivative works reimagining canonical stories, proliferates online, fostering community-driven narratives often centered on marginalized perspectives. Web serials, episodic stories published chapter-by-chapter on platforms like since the 2010s, have boomed, enabling aspiring authors to build audiences directly and leading to traditional publications, as seen in the platform's role in amplifying diverse voices through over 400 million user-generated stories as of 2020. Key impacts of literary text include and adaptation challenges, exemplified by banned works that provoke societal debate. Vladimir Nabokov's (1955), a controversial exploring obsession through unreliable narration, faced bans in the UK (1955–1959), (1956–1958), and other countries for alleged , yet its literary merit—praised for stylistic innovation—ultimately secured its place in the , underscoring text's power to confront taboos.

Visual and Performing Arts

In , text has served as both a medium and a subject, transforming into visual forms that challenge traditional boundaries between literature and imagery. , pioneered by in his 1918 collection , arranges words and letters on the page to form visual shapes that echo the poem's content, such as raindrops or Eiffel Tower silhouettes, emphasizing the materiality of text itself. Similarly, conceptual artist On Kawara's "Today" series, begun in 1966, consists of minimalist paintings of dates in white acrylic on monochrome grounds, reducing text to its essential temporal function and inviting viewers to confront the passage of time. Text-based installations further exemplify this integration, as seen in Jenny Holzer's Truisms (1977), where aphoristic phrases like "Abuse of power comes as no surprise" were projected or displayed on LED signs in public spaces, blending text with architecture to provoke social reflection. Graffiti and have elevated ephemeral text into enduring public statements, with Banksy's works from the late onward incorporating stenciled words like "Keep it real" alongside imagery to critique and , often appearing unannounced on urban walls. , a technique rooted in 7th-century Islamic traditions, treats script as ornamental art, where and Naskh styles adorn manuscripts and architecture, as in the Dome of the Rock's Quranic inscriptions (691–692 ), prioritizing aesthetic harmony over legibility. In digital contexts, emerged in the 1960s using characters to create images, such as early computer-generated banners, prefiguring pixelated text visuals in modern aesthetics. Recent innovations extend text's visual role into interactive and virtual realms. Augmented reality (AR) overlays, popularized post-2010 through apps like Pokémon GO (2016), integrate dynamic text elements into physical environments for artistic installations, such as Rafael Lozano-Hemmer's Pulse Room (2006, AR-enhanced versions post-2010s), where heart-rate data appears as floating words. The 2021 NFT boom introduced text-centric digital art, with projects like Pak's The Merge (2021) using blockchain-inscribed phrases as collectible visuals, selling for millions and redefining text's scarcity and ownership. By 2023–2025, AI-generated visual text has surged, with tools like Midjourney and DALL-E creating surreal typographic landscapes—e.g., Salvador Dalí-inspired melting letters—enabling artists to explore generative narratives without manual rendering. In , text functions as foundational script and ephemeral enhancement, guiding narrative delivery while interacting with movement and sound. Play scripts, formalized in printed editions like the of Shakespeare's works (1623), compile dialogues and stage directions to standardize performances, influencing theatrical traditions from Elizabethan stages to contemporary revivals. Operatic librettos, originating in 17th-century with Claudio Monteverdi's Orfeo (1607), provide sung text that drives dramatic action, often adapting literary sources into verse for musical integration. , standardized in the 1920s for silent films and theater accessibility—first notably used in (1927) with intertitles—evolved into real-time captions, enhancing global reach without disrupting performative flow. These elements underscore text's performative vitality, where words are not merely read but embodied and synchronized with visual and auditory cues.

Media and Communication

In , text serves as the primary medium for conveying factual information efficiently, with the inverted pyramid style emerging in the late as a response to telegraph limitations, prioritizing the most essential details at the outset followed by supporting facts. This structure, which originated between 1880 and 1890 amid technological and commercial pressures on news dissemination, allows editors to truncate stories without losing core meaning, a practice that became standard in print and persists in digital formats. In the post-2010s digital era, standards have evolved to combat , with organizations like the International Fact-Checking Network establishing guidelines for transparency, sourcing, and bias avoidance in verifying claims, with broader counts such as the Duke Reporters' Lab's tally of over 440 active fact-checking projects worldwide by 2025 that reduce false beliefs across countries. These standards emphasize rigorous of viral content, drawing on epistemological frameworks to evaluate political and societal claims. Advertising leverages concise, persuasive text to influence consumer behavior, exemplified by Nike's "Just Do It" slogan introduced in 1988 by ad agency , which encapsulated motivational and boosted the brand's through its universal in campaigns across billboards and early media. techniques often employ Aristotle's —ethos for credibility, pathos for emotional appeal, and logos for logical arguments—to craft slogans and posts that resonate on , where short-form text like captions or overlays drives engagement by evoking urgency or aspiration. Billboards, in particular, use bold, imperative phrasing to capture fleeting attention, integrating rhetorical devices such as and to reinforce brand messages amid high-traffic environments. Mass communication utilizes text for broad societal influence, as seen in World War II propaganda posters where succinct slogans like "Loose Lips Sink Ships" combined with visual elements to promote vigilance and recruitment, fostering national unity through emotional and patriotic appeals produced by government offices. In the digital age, social media micro-text has transformed this landscape since 's launch in 2006 with a 140-character limit inspired by constraints, expanded to 280 characters in 2017 to allow more expressive yet concise messaging that amplifies public discourse. Platforms like (now X) enable rapid dissemination of ideas, but their brevity often condenses complex narratives into viral snippets, shaping global conversations on events from to social movements. Ethical considerations in text-based media highlight the risks of , particularly in the with AI-generated deepfakes and synthetic text that fabricate credible narratives, as evidenced by manipulated content during elections that erodes trust and influences public opinion, with studies showing prior exposure increases susceptibility across demographics. Accessibility laws, such as the Americans with Disabilities Act of , mandate captioned media for deaf and hard-of-hearing individuals, requiring synchronized text overlays in videos to ensure equitable communication under Title III provisions for public accommodations. Societal impacts extend to algorithms' text moderation, where regulations like the EU's (effective 2024) impose transparency and risk assessments on platforms to curb harmful content, including and , with fines up to 6% of global revenue for non-compliance. In the Global South, digital divides exacerbate these issues, as limited and hinder participation in text-based , widening inequalities in and , with nearly 2.7 billion people offline as of 2023 primarily in low-income regions.

References

  1. [1]
    TEXT Definition & Meaning - Merriam-Webster
    1. a (1) : the original words and form of a written or printed work (2) : an edited or emended copy of an original work b : a work containing such text.Synonyms of textText messageProof textCover textPlaintext
  2. [2]
    Definition and Examples of Text in Language Studies - ThoughtCo
    Jul 3, 2019 · Text linguistics refers to a form of discourse analysis—a method of studying written or spoken language—that is concerned with the description ...
  3. [3]
    [PDF] Context and Text - Academy Publication
    A text is a unit of language in use. It is not a grammatical unit, like a clause or a sentence; and it is not defined by its size.
  4. [4]
    Cohesion in English | M.A.K. Halliday, Ruqaiya Hasan
    Jan 14, 2014 · A principal component of these resources is 'cohesion'. This book studies the cohesion that arises from semantic relations between sentences.
  5. [5]
    Definition and Examples of Text Linguistics - ThoughtCo
    May 1, 2025 · Text linguistics is a branch of linguistics concerned with the description and analysis of extended texts (either spoken or written) in communicative contexts.
  6. [6]
    What Is Text in Computing?
    Jul 9, 2025 · 1. In general, text is a collection of words or letters that are understandable by the reader. On a computer, text is added, viewed, edited ...
  7. [7]
    Definition of text | PCMag
    Words, sentences and paragraphs (precisely what you are reading). A page of text takes approximately from 2,000 to 4,000 bytes of storage.
  8. [8]
    What is text, really? | Journal of Computing in Higher Education
    The authors argue that text is best represented as an ordered hierarchy of content object (OHCO), because that is what text really is.
  9. [9]
    What is a Text? – Advanced English - BC Open Textbooks
    A text is anything that conveys a set of meanings to the person who examines it. You might have thought that texts were limited to written materials.
  10. [10]
    Text - Purdue OWL
    “Text” refers to a written or typed document. In terms of a rhetorical situation, however, “text” means any form of communication that humans create.
  11. [11]
    Semiotics: Sign, Signifier, Signified | Writing Commons
    Semiotics is a theory of communication. Learn how humans use & interpret signs to communicate, to learn, and to develop knowledge.What is Semiotics? · History · Ferdinand de Saussure · Charles Sanders Peirce
  12. [12]
    Peirce's theory of signs - Stanford Encyclopedia of Philosophy
    Oct 13, 2006 · Peirce's Sign Theory, or Semiotic, is an account of signification, representation, reference and meaning. Although sign theories have a long ...<|separator|>
  13. [13]
    Recommendation concerning the preservation of, and access to
    Nov 17, 2015 · It is preservable and usually moveable. The content may comprise signs or codes (such as text), images (still or moving) and sounds, which can ...
  14. [14]
    The Evolution of Writing | Denise Schmandt-Besserat
    Feb 6, 2021 · The evolution of writing from tokens to pictography, syllabary and alphabet illustrates the development of information processing to deal with larger amounts ...Introduction · 2. Pictography: Writing As... · 3. Logography: Shift From...Missing: hypertext | Show results with:hypertext<|separator|>
  15. [15]
    On the Notion of Text and Its Boundaries in the Context of Semiotics ...
    Aug 9, 2025 · To sum up, I define the text as a semiotic conditional reality generated by a communicative situation with non-degenerated planes of expression ...
  16. [16]
    [PDF] a comparison of the effect of textual, audio and - ERIC
    Audio is a different sensory modality from visual modality, such as printed text and pictures, because audio is processed by the auditory working memory ...
  17. [17]
    Language awareness: On the semiotics of talk and text - ScienceDirect
    Language awareness involves recognizing that talk and text are different semiotic phenomena, with talk being dynamic and text static, and that viewing written ...
  18. [18]
    Text - Etymology, Origin & Meaning
    Originating from Latin textus meaning "thing woven," text refers to written words or scriptures and also to sending messages via mobile systems since 2005.
  19. [19]
    text, n.¹ meanings, etymology and more | Oxford English Dictionary
    The earliest known use of the noun text is in the Middle English period (1150—1500). OED's earliest evidence for text is from before 1387, in a translation by ...Missing: evolution | Show results with:evolution
  20. [20]
    The Gutenberg Press - Oregon State University Special Collections
    In 1436 Johaness Gutenberg, a German goldsmith, began designing a machine capable of producing pages of text at an incredible speed.
  21. [21]
    Johannes Gutenberg | Printing Press, Inventions, Facts ... - Britannica
    Oct 10, 2025 · Gutenberg's printing press was considered a history-changing invention, making books widely accessible and ushering in an “information ...
  22. [22]
    History of Hypertext: Article by Jakob Nielsen - NN/G
    Feb 1, 1995 · Xanadu (1965) The actual word "hypertext" was coined by Ted Nelson in 1965. Nelson was an early hypertext pioneer with his Xanadu system, which ...
  23. [23]
    Emoji: The Complete History - WIRED
    Feb 1, 2018 · Emoji have been available outside of Japan since the mid-2000s through separate apps, which let users copy and paste the icons into text ...Missing: extension | Show results with:extension
  24. [24]
  25. [25]
    How Egyptian hieroglyphs were decoded, a timeline to decipherment
    Hieroglyphs, meaning 'sacred carving', were largely carved into stone and commonly used in temples, tombs and on other monuments for nearly 3000 years.
  26. [26]
    Egyptian Symbols and Figures: Hieroglyphs | NEH-Edsitement
    The ancient Egyptians created a form of picture-writing known as hieroglyphs around 3100 BCE. Each picture was a symbol representing something they observed in ...Missing: origins | Show results with:origins
  27. [27]
    [PDF] Dating the Origin of Chinese Writing: Evidence from Oracle Bone ...
    The origin of Chinese writing is estimated to be no later than 2100 BCE, with the earliest available writing from late Shang (ca. 1300-1046 BCE).
  28. [28]
    Reading the Calendar Glyphs | Living Maya Time
    Since the 1950s, scholars from all over the world have been able to decipher and read a large portion of the Maya hieroglyphs. These hieroglyphs are found ...
  29. [29]
    The Origins and Evolution of Alphabet Letter Sounds Through ...
    Number of symbols in Phoenician alphabet, 22 consonant characters, no vowels (circa 1050 BCE). Greek vowels introduced ; Age of Latin alphabet standardization ...
  30. [30]
    Papyrus: A Brief History – Dartmouth Ancient Books Lab
    May 23, 2016 · Around 3000 BC, the Egyptians would revolutionize the literary world by producing a smooth, flexible writing material that could accept and ...
  31. [31]
    Scroll and Codex - Sir Thomas Browne
    Stitched together and protected by a cover, the parchment notebook was used for accounts, notes, drafts, and letters. The earliest evidence for its literary use ...Missing: 1st | Show results with:1st
  32. [32]
    [PDF] Evolution of the Chinese book - Northeastern repository
    Bi Sheng (970 – 1051) invented clay types in the. Northern Song Dynasty ... The invention of movable type printing was also a major breakthrough in Chinese book ...
  33. [33]
    The History of Printing in Asia According to Library of Congress ...
    Jun 22, 2021 · In this view, moveable type printing traveled west from China along the Silk Road to the Uighurs at Turfan in the 11th century.
  34. [34]
    [PDF] An Empirical Classification of Civilizations Based on Writing Systems
    Oct 1, 2003 · The overland Silk Roads diffused Aramaic forms to the east, and these developed into the Central Asian systems of Sogdian, Uighur, Kok. Turki ...
  35. [35]
    [PDF] Between Text and Language: Unicode and the Rise of Emojis
    Dec 14, 2021 · It was developed during the late 1980s as a replacement for previous-generation encoding standards such as ASCII and EBCDIC, which limited the ...Missing: precursors | Show results with:precursors
  36. [36]
    Understanding typography - Material Design 2
    Typography expresses hierarchy and brand presence. A typeface is a collection of letters with shared patterns. Type aligns to a 4dp baseline grid.
  37. [37]
    Top 10 Principles of Typography in Design for Stunning Visuals
    Aug 14, 2024 · Serif Typefaces: Suitable for printed books and formal documents where tradition and readability are key. Sans-Serif Typefaces: Ideal for ...
  38. [38]
    Kerning, Leading, and Tracking in Typography Design: Differences ...
    Nov 13, 2024 · Kerning is the space between letters in a word, leading is the vertical space between lines, and tracking is the uniform spacing across a word ...
  39. [39]
    Typography Basics - UX Collective
    Oct 24, 2025 · Serif and sans-serif typefaces differ in tone partly because of these structures: serifs guide the eye along lines of text, while sans-serifs ...Missing: core | Show results with:core
  40. [40]
    Type and Typography - PrintWiki
    Initially, printers strove to duplicate the style of lettering used in manuscripts—the black letter, or gothic style found in Gutenberg's Bible is one ...
  41. [41]
    Understanding typography: Giambattista Bodoni and the invention of ...
    Jan 25, 2023 · Bodoni inflicted in his letterforms the crisp contrast and sharp edges that resulted from an engraving burin, reflecting a style that was ...
  42. [42]
    [PDF] Making a Hybrid of Fraktur and Helvetica - DiVA portal
    With. Helvetica as a Swiss invention and Fraktur as a symbol for Germany, the names of the typeface and weights are a tribute to the two and their history. With ...
  43. [43]
    Jan Tschichold: Defining the New Typography Movement
    May 24, 2025 · He championed the New Typography movement, emphasizing asymmetry, sans-serif fonts, and clear hierarchy. Later, he shifted to more classical ...
  44. [44]
    [PDF] The Principles of the New Typography - Graphic Design Reading
    In small items of printed matter, 12 to 24 points are the minimum margin required; in posters 48 points. On the other hand, borders of solid red or black can be ...Missing: hierarchy grid
  45. [45]
    History and Evolution of Typography - GeeksforGeeks
    May 14, 2024 · The early printed books, known as incunabula, were using blackletter and roman typefaces. However, as printing technology evolved new typeface ...
  46. [46]
    [PDF] The Designers' Guide to Professional Typography
    Giambattista Bodoni, Bodoni. William Caslon, Caslon. Firmin Didot, Didot ... takes a high degree of skill in PostScript drawing tools, as well as a highly.
  47. [47]
    Accessible typography - APA Style
    For example, a person living with severe dyslexia could choose to have the font swapped in real time with a typeface and spacing that works better for them— ...Myth 1: Serif Fonts Are Not... · Myth 3: Smart Quotes Are Not... · Typeface StrategiesMissing: dashes | Show results with:dashes
  48. [48]
    [PDF] BDA Dyslexia Style Guide
    Alternatives include Verdana, Tahoma, Century Gothic, Trebuchet. •. Font size should be 12-14 point. Some dyslexic readers may request a larger font.Missing: em | Show results with:em
  49. [49]
    Variable Fonts on the Web - WebKit
    Nov 2, 2016 · Variable fonts in OpenType allow multiple axes to control font aspects, using floating-point values, and can improve loading by reducing ...
  50. [50]
    CSS Fonts Module Level 4 - W3C
    Feb 1, 2024 · This specification defines modifications to the existing CSS Fonts 3 specification along with additional features.
  51. [51]
    The Complete Guide To Sustainable Font Choices In Digital Design ...
    Sep 4, 2024 · In print, the use of eco fonts has been documented to substantial ink savings, sometimes up to 28% compared to traditional fonts. This not only ...Missing: 2020s | Show results with:2020s
  52. [52]
    11.8: Introduction to Linguistics - Social Sci LibreTexts
    Jun 18, 2021 · These include phonetics, phonology, morphology, syntax, semantics, and pragmatics.
  53. [53]
    5 Domains of Language: Best of Therapy Tools! February 2021
    Mar 15, 2021 · The 5 domains of language include: phonology, morphology, syntax, semantics, and pragmatics. Why are they important for therapy and language development?
  54. [54]
    [PDF] Introduction to Linguistics
    Grammars are ways to generate signs from more basic signs. Signs combine a form and a meaning, and they are identical with neither their exponent nor with their ...
  55. [55]
    [PDF] Semantics, Pragmatics and Discourse
    Aug 22, 2011 · The dynamic semantic point of view: the text + its context is the unit of meaning. Each sentence is interpreted relative to a context and in ...
  56. [56]
    (PDF) The Semantics and Pragmatics of Polysemy: A Relevance
    This thesis investigates the phenomenon of polysemy: a single lexical form with two or multiple related senses (e.g. catch the rabbit/order the rabbit; ...
  57. [57]
    [PDF] Key Terms in Discourse Analysis | uogbooks
    in three ways – as semantics, phonology and lexico-grammar – and text can be analysed with respect to three meta-functions: ideational, interpersonal and ...
  58. [58]
    (PDF) A Synopsis of the Lexical Variations in British and American ...
    May 6, 2025 · The focal point of this research work is to find out the lexical distinctions between British and American varieties of English and their persistently ...
  59. [59]
    (PDF) Code-Switching in Multilingual Societies - ResearchGate
    Aug 6, 2025 · The study revealed how code-switching served as a versatile linguistic strategy, accommodating different communicative needs and reflecting speakers' social ...
  60. [60]
    From DDoS to Delulu: How the Language of the Internet Has Evolved
    Apr 16, 2025 · Other acronyms gained traction in the 1990s, including “TMI” (too much information) and “G2G” (got to go). TMI” became particularly popular ...
  61. [61]
    Philosophy of Linguistics
    Sep 21, 2011 · The influential Swiss linguist Ferdinand de Saussure (1916) distinguished between langue, a socially shared set of abstract conventions ...
  62. [62]
    A Brief Analysis of Chomsky's Transformational Generative Grammar
    In this study, we interpreted Chomsky's transformational generative grammar, and further elaborated the significance and influence of transformational ...
  63. [63]
    Computational Linguistics - Stanford Encyclopedia of Philosophy
    Feb 6, 2014 · Computational linguistics is the scientific and engineering discipline concerned with understanding written and spoken language from a computational ...
  64. [64]
    [PDF] A perishable ability? The future of writing in the face of generative ...
    Aug 26, 2025 · The 2020s have been witnessing a very significant advance in the development of generative artificial intelligence tools, including text ...
  65. [65]
  66. [66]
    Milestones:American Standard Code for Information Interchange ...
    May 23, 2025 · The American Standards Association X3.2 subcommittee published the first edition of the ASCII standard in 1963. Its first widespread commercial ...
  67. [67]
    ISO8859 family - IBM
    Each code set includes the ASCII character set plus its own unique character set. The ISO8859 encoding figure shows the ISO8859 general encoding scheme.<|separator|>
  68. [68]
    [PDF] The Unicode Standard, Version 16.0 – Core Specification
    Sep 10, 2024 · Starting with this version, the Unicode Consortium has changed the way the Unicode Standard is produced. The interactive HTML version is ...
  69. [69]
    Usage statistics of UTF-8 for websites - W3Techs
    UTF-8 is used by 98.8% of all websites whose character encoding is known. Popular sites using it include Google.com, Microsoft.com, and Facebook.com.
  70. [70]
    Using Unicode Normalization to Represent Strings - Win32 apps
    Jan 7, 2021 · To use Unicode normalization, an application can call the NormalizeString and IsNormalizedString functions for rearrangement of strings acccording to Unicode 4 ...
  71. [71]
  72. [72]
    The IBM punched card
    Data was assigned to the card by punching holes, each one representing a character, in each column. When writing a program, one card represented a line of code ...
  73. [73]
  74. [74]
    Supervised Quantum Machine Learning: A Future Outlook ... - arXiv
    Jun 17, 2025 · The goal was to explore whether shallow quantum circuits could process text representations in a way that yields meaningful learning performance ...
  75. [75]
    Tokenization - Stanford NLP Group
    Tokenization is the task of chopping it up into pieces, called tokens, perhaps at the same time throwing away certain characters, such as punctuation.
  76. [76]
    Ten years of Google Translate - The Keyword
    Apr 28, 2016 · In 2011, we first introduced the ability to have a bilingual conversation on Google Translate. The app will recognize which language is being ...
  77. [77]
    Regular Expression Matching Can Be Simple And Fast
    This article reviews the good theory: regular expressions, finite automata, and a regular expression search algorithm invented by Ken Thompson in the mid-1960s.<|control11|><|separator|>
  78. [78]
    NLTK: the natural language toolkit - ACM Digital Library
    NLTK is written in Python and distributed under the GPL open source license. ... date: 28-Sep-2025. https://dl.acm.org/doi/10.1145/3746059.3747600. Wang ZLin S ...
  79. [79]
    LLM ethics benchmark: a three-dimensional assessment system for ...
    Oct 5, 2025 · Published: 05 October 2025. LLM ethics benchmark: a three-dimensional assessment system for evaluating moral reasoning in large language models.
  80. [80]
    The Epic of Gilgamesh - Yale University Press
    Apr 30, 2020 · It was composed nearly 4,000 years ago in ancient Mesopotamia (roughly equivalent to where Iraq and eastern Syria are now). No one knows who ...
  81. [81]
    Guide to the classics: Don Quixote, the world's first modern novel
    Jun 25, 2018 · Don Quixote of La Mancha by Miguel de Cervantes, the first modern novel. Published in two parts in 1605 and 1615, this is the story of Alonso Quijano.
  82. [82]
    Freytag's Pyramid: Definition, Elements and Example - MasterClass
    Feb 15, 2023 · Gustav Freytag, the novelist behind Freytag's Pyramid, created a five-act plot structure, borrowing from Aristotle and Horace.
  83. [83]
    Stream of consciousness as a narrative technique in the novel Ulysses
    Stream of consciousness is a narrative technique that depicts the continuous flow of thoughts and sensations experienced by characters' minds.Missing: Freytag's pyramid
  84. [84]
    A Thousand and One Nights: Arabian Story-telling in World Literature
    Oct 26, 2017 · "A Thousand and One Nights" is a collection of stories, a "frame story" where Shahrazad tells tales to King Shahrayar, and has influenced  ...Missing: roles canon Western
  85. [85]
    [PDF] Challenges and Possibilities for World Literature ... - Purdue e-Pubs
    David Bellos, who has studied the UNESCO statistics for the translation of literary works between seven languages between 2000 and 2009, comes to the striking ...
  86. [86]
    From Infocom to 80 Days: An oral history of text games and ...
    Jun 20, 2024 · As computers spread to more people around the world, interactive fiction was able to reach more and more readers. Too good to be grue.Missing: serials Wattpad
  87. [87]
    Fanfiction, Digital Platforms, and Social Reading (Chapter 9)
    Nov 29, 2024 · It has long been argued that digital textuality fundamentally alters familiar conceptions of literary authorship. Critics such as Jay David ...
  88. [88]
    Wattpad as a resource for literary studies. Quantitative and ...
    We present several analyses of how fiction is transmitted through the social reading platform Wattpad, one of the largest platforms for user-generated stories.