Knowledge, Technology, and Power
Welcome, Knowledge Explorers!
Welcome to one of the most urgent chapters in our entire course. Every time you search the internet, scroll through a social media feed, or ask a chatbot a question, invisible systems are shaping what knowledge reaches you — and what gets filtered out. But how do we know whether the information we receive reflects reality, or merely reflects the priorities of those who built the systems delivering it? Let's explore how technology and power intersect to shape the knowledge landscape of our time.
Tools in Knowledge
Human beings have always used tools to extend their capacity to produce, record, and share knowledge. A tool in knowledge is any instrument, technique, or system that enhances our ability to observe, measure, analyze, or communicate. The telescope allowed Galileo to observe the moons of Jupiter. The printing press enabled the rapid dissemination of ideas across Europe. Each of these tools did not merely assist knowledge production — they transformed it, opening entirely new domains of inquiry.
Today, digital technologies represent the most powerful knowledge tools ever created. Computers process billions of calculations per second. Sensors detect phenomena far beyond human perception. Networks transmit information globally in milliseconds. But here is the crucial epistemological question: do these tools simply help us see more clearly, or do they actively shape what we see and how we understand it? The philosopher Marshall McLuhan argued that "the medium is the message" — the tools we use to communicate fundamentally alter the nature of the information itself.
Digital Knowledge Systems
A digital knowledge system is any technology-based platform or infrastructure that organizes, stores, retrieves, or distributes knowledge — including search engines, online encyclopedias, academic databases, and social media platforms. Consider the transformation: a student in the 1980s researching climate change would visit a physical library and search a card catalog. Today, that student can access millions of documents in seconds from a smartphone. This represents extraordinary democratization — but also introduces new epistemological challenges.
| Feature | Traditional Knowledge Systems | Digital Knowledge Systems |
|---|---|---|
| Access speed | Hours to weeks | Milliseconds to seconds |
| Geographic reach | Local or regional | Global |
| Gatekeepers | Librarians, editors, publishers | Algorithms, platform designers |
| Verification | Peer review, editorial boards | Variable — often minimal |
| Persistence | Physical degradation over time | Can be deleted or altered instantly |
| Volume | Limited by physical space | Virtually unlimited |
As the table shows, digital knowledge systems offer tremendous advantages in speed, reach, and volume. But they also shift who controls access — from human gatekeepers with visible criteria to algorithmic systems whose criteria may be hidden. This shift has profound implications for how knowledge is produced, validated, and distributed.
Algorithms
An algorithm is a set of step-by-step instructions that a computer follows to perform a task or solve a problem. Algorithms determine what appears in your search results, which posts you see on social media, what products are recommended to you, and even what news stories are presented as important. They are the invisible architects of your digital information environment.
Search engine algorithms do not simply find all relevant pages and present them neutrally. They rank results according to complex criteria — relevance, popularity, recency, and the searcher's past behavior. Two people searching for the same term may receive very different results. This raises a fundamental epistemological question: when an algorithm selects and ranks information for you, is it helping you find knowledge, or is it constructing a particular version of knowledge that serves certain interests?
Diagram: How Algorithms Filter Knowledge
How Algorithms Filter Knowledge
Type: diagram
sim-id: algorithm-filter-knowledge
Library: p5.js
Status: Specified
Bloom Level: Analyze (L4) Bloom Verb: Trace Learning Objective: Trace how algorithmic filtering transforms raw information into the curated knowledge a user actually sees.
Instructional Rationale: A flow diagram showing information passing through successive algorithmic filters makes visible the normally invisible process of algorithmic curation, helping students understand that search results and feeds are constructed, not neutral.
Visual elements: - Left side: A large pool labeled "All Available Information" with many colored dots representing diverse sources - Center: A series of filter layers labeled "Relevance Filter," "Popularity Filter," "Personalization Filter," and "Monetization Filter" - Right side: A small curated set labeled "What You See" with fewer dots - Arrows showing information flow from left to right, with dots being removed at each filter stage - A counter showing how many items are filtered out at each stage
Interactive controls: - Slider to adjust the "strictness" of each filter and see how results change - Toggle to enable/disable each filter layer independently - Button to randomize the initial information pool - Hover over filtered-out dots to see why they were removed
Default state: All four filters active, showing dramatic reduction from hundreds of items to a small handful.
Color scheme: Information dots in varied colors representing different sources/perspectives, filters in graduated teal shades, final results highlighted in amber.
Responsive design: Canvas resizes to fit container width. Flow diagram shifts from horizontal to vertical on narrow screens.
Implementation: p5.js with createSlider() for filter strictness and createCheckbox() for filter toggles
Big Data
Big data refers to datasets so large and complex that traditional methods of analysis cannot process them effectively. Big data is characterized by three properties often called the "three Vs": volume (enormous quantities), velocity (high-speed collection), and variety (many formats — text, images, sensor readings, transactions).
Big data has transformed knowledge production across fields. Epidemiologists track disease patterns across millions of health records. Climate scientists model atmospheric systems using petabytes of satellite data. In each case, big data reveals patterns invisible in smaller datasets.
But big data also raises serious epistemological concerns. A correlation in a massive dataset is not the same as a causal explanation. The sheer volume can create a false sense of certainty. Yet data always reflects the conditions of its collection — if certain populations are underrepresented, the patterns extracted will be systematically skewed. The old programming adage — "garbage in, garbage out" — applies with full force.
Sofia's Reflection
Here is something worth pausing over: more data does not automatically mean better knowledge. A dataset of ten million medical records sounds impressive — but if those records come overwhelmingly from one demographic group, the patterns you find may not apply to everyone. The size of the dataset can actually mask its biases, making flawed conclusions look more credible than they really are. What questions should we ask about any dataset before trusting the knowledge it produces?
Machine Learning
Machine learning is a branch of artificial intelligence in which computer systems improve their performance by learning from data, rather than being explicitly programmed with rules. Instead of a programmer writing "if the email contains these words, mark it as spam," a machine learning system is given thousands of examples and learns to identify patterns on its own.
Machine learning systems now make consequential decisions in medical diagnosis, criminal sentencing, loan approvals, and hiring. The epistemological challenge is clear: these systems produce outputs but often cannot explain why they reached a particular conclusion. This is the "black box" problem — a model might accurately predict which patients are at risk for heart disease, but doctors may not understand which factors the model considers most important.
| Aspect | Traditional Programming | Machine Learning |
|---|---|---|
| How rules are created | Written by programmers | Learned from data |
| Transparency | Rules can be inspected | Often opaque ("black box") |
| Adaptability | Fixed unless reprogrammed | Improves with more data |
| Bias source | Programmer assumptions | Training data composition |
| Explanation | Can explain each decision | Often cannot explain decisions |
Artificial Intelligence
Artificial intelligence (AI) is the broad field of creating computer systems capable of performing tasks that typically require human intelligence — understanding language, recognizing images, making decisions, and generating creative content. Machine learning is one approach within AI, but the field also includes rule-based systems, robotics, and natural language processing.
AI systems are now deeply embedded in everyday knowledge practices. Language models generate text and translate between languages. Image recognition systems classify photographs and detect faces. Each application involves AI making judgments about meaning, relevance, and quality — judgments previously the exclusive domain of human knowers.
The philosophical questions are profound. Can a machine truly "understand" language, or does it merely manipulate symbols? When an AI generates a plausible answer, does it "know" the answer? An AI system does not believe anything, does not care whether its outputs are true, and does not justify its claims through reasoning. Yet its outputs can be extraordinarily useful and often indistinguishable from human-generated knowledge.
Diagram: Mapping AI in Knowledge Production
Mapping AI in Knowledge Production
Type: concept map
sim-id: ai-knowledge-production-map
Library: p5.js
Status: Specified
Bloom Level: Understand (L2) Bloom Verb: Classify Learning Objective: Classify the different roles AI plays across stages of knowledge production, from data collection to dissemination.
Instructional Rationale: A concept map showing AI's involvement at each stage of the knowledge production pipeline helps students see that AI is not a single phenomenon but a collection of tools operating at many points in the epistemic process.
Visual elements: - A horizontal pipeline with stages: "Data Collection," "Data Processing," "Analysis," "Interpretation," "Dissemination" - At each stage, branching nodes show specific AI applications (e.g., under Data Collection: "Sensors," "Web Scraping," "Satellite Imaging") - Color-coded nodes indicating level of human oversight: green (high oversight), yellow (moderate), red (minimal) - Dotted lines connecting related applications across stages
Interactive controls: - Click on any node to see a brief description and real-world example - Filter toggle to show only applications in a specific Area of Knowledge (Natural Sciences, Human Sciences, Arts) - Slider to adjust "level of human oversight" threshold, dimming nodes below the threshold
Default state: Full pipeline visible with all nodes shown, colored by oversight level.
Color scheme: Pipeline stages in neutral gray, nodes colored by oversight level (teal for high, amber for moderate, coral for minimal).
Responsive design: Canvas resizes to fit container width. Pipeline shifts from horizontal to vertical on narrow screens.
Implementation: p5.js with clickable regions and createSelect() for AOK filter
Algorithmic Bias
Algorithmic bias occurs when a computer system produces results that are systematically unfair to certain groups. This bias typically enters through training data — if a system learns from data reflecting existing social inequalities, it will reproduce those inequalities.
The cases are sobering. Facial recognition systems misidentify people with darker skin tones at significantly higher rates — because training datasets contained far more light-skinned faces. Hiring algorithms have discriminated against women because the historical data reflected decades of gender discrimination. These are not merely technical glitches. They are epistemological problems rooted in the data that shapes what these systems "know." When deployed at scale, biased algorithms do not just reflect inequalities — they amplify and entrench them, giving systemic bias the appearance of data-driven neutrality.
Watch Out!
Be careful not to assume that algorithmic decisions are objective simply because they are made by computers. The phrase "data-driven" can create a false sense of neutrality. Every dataset reflects choices about what to measure, whom to include, and how to categorize — and those choices embed human values and biases. When someone claims a decision is "just what the algorithm says," always ask: what data was it trained on, and whose experiences are missing from that data?
Data Visualization
Data visualization is the representation of data in graphical form — charts, graphs, maps, and interactive displays. A well-designed graph can reveal patterns invisible in raw numbers. But visualization is also a tool of persuasion. A designer's choices — what scale to use, where to start the y-axis, which data to include — shape how viewers interpret the data. A bar chart with a truncated y-axis can make a small difference look dramatic.
This means that digital literacy — the ability to critically evaluate digital information and its systems — must include visual literacy. When you encounter a visualization, ask: What data is shown, and what is left out? Does the representation accurately reflect the underlying data, or does it exaggerate certain features?
Simulations
A simulation is a computer model that represents a real-world system, allowing researchers to experiment with it virtually. Climate scientists simulate atmospheric conditions decades ahead. Epidemiologists simulate disease spread. Engineers simulate structural integrity before construction begins.
Simulations are extraordinary because they let us study systems too large, too slow, or too dangerous to experiment with directly. You cannot deliberately cause a pandemic to study its spread, but you can simulate one. However, every simulation rests on assumptions. A climate model must decide which variables to include and what resolution to use. The knowledge produced is always conditional — it tells us what would happen if the model's assumptions are correct. The map is not the territory.
Diagram: How Simulations Generate Knowledge
How Simulations Generate Knowledge
Type: diagram
sim-id: simulation-knowledge-cycle
Library: p5.js
Status: Specified
Bloom Level: Evaluate (L5) Bloom Verb: Assess Learning Objective: Assess the strengths and limitations of simulation-based knowledge by tracing the cycle from assumptions to outputs and validation.
Instructional Rationale: A cyclical diagram showing how simulations work — from assumptions through modeling to outputs and back to validation — helps students evaluate rather than uncritically accept simulation results.
Visual elements: - A circular flow with four major stages: "Assumptions & Parameters," "Model Building," "Simulation Output," and "Validation Against Reality" - At each stage, cards listing key questions (e.g., at Assumptions: "What variables are included?", "What is simplified or excluded?") - A "confidence meter" that rises or falls as students toggle different validation outcomes - A side panel showing a concrete example (e.g., a simple epidemic simulation)
Interactive controls: - Click on each stage to expand key questions and examples - Toggle switches for "Model includes variable X" to see how adding/removing assumptions changes output - A "Run Simulation" button that animates the cycle with a simple SIR epidemic model - Slider for initial parameters (e.g., transmission rate) to show sensitivity to assumptions
Default state: Full cycle visible with the epidemic example loaded, confidence meter at moderate level.
Color scheme: Stages in graduated teal, question cards in cream, confidence meter in amber-to-green gradient.
Responsive design: Canvas resizes to fit container width. Cycle diagram adjusts radius on narrow screens.
Implementation: p5.js with createButton() for run control, createSlider() for parameters, and createCheckbox() for variable toggles
Power and Knowledge
The relationship between power and knowledge was articulated most influentially by the French philosopher Michel Foucault, who argued that the two are deeply intertwined. Those who hold power shape what counts as knowledge, and those who control knowledge gain power.
Consider how this works. Governments fund certain research programs and not others. Corporations control vast datasets but restrict access. Media companies decide which stories receive attention. Universities determine which disciplines are prestigious. In each case, power structures shape the knowledge landscape — determining what questions are asked, what methods are legitimate, and whose voices are heard. This is not a conspiracy theory but a structural observation. The question is not whether power influences knowledge but how, and whether those influences are visible and accountable.
Gatekeeping
Gatekeeping is the process by which individuals, institutions, or systems control access to information, platforms, or audiences. In traditional media, editors and publishers served as gatekeepers. In academia, peer reviewers perform this function. In the digital age, platform algorithms and content moderation policies are the new gatekeepers.
Gatekeeping serves legitimate functions — peer review ensures rigor, editorial judgment supports accuracy, content moderation prevents dangerous misinformation. But when gatekeeping power is concentrated in few hands, it can suppress legitimate knowledge and silence dissenting voices. The epistemological challenge is to design gatekeeping systems that maintain quality without systematically excluding certain perspectives.
Politics of Expertise
The politics of expertise refers to how societies decide who counts as an expert and whose expertise is valued. In practice, who gets recognized as an expert is shaped by education, institutional affiliation, race, gender, and cultural background — not only by knowledge and competence.
Consider a climate scientist with decades of research and a social media influencer with millions of followers. Both make claims about climate change, but their authority differs fundamentally. The scientist's authority rests on methodology and peer review. The influencer's rests on popularity and platform reach. When these two forms of authority compete, societies struggle to make informed decisions about complex issues — making the politics of expertise an urgent practical challenge.
Trust in Institutions
Trust in institutions — universities, scientific organizations, government agencies, media outlets, courts — is foundational to shared knowledge. Most of what any individual "knows" comes not from personal experience but from institutional knowledge systems.
When institutional trust erodes, consequences are severe: declining vaccination rates, undermined governance, impossible evidence-based policy. But institutional trust is not blind faith — it is earned through transparency, accountability, and reliability. The epistemological challenge is maintaining appropriate trust: neither naive acceptance nor cynical rejection.
Sofia's Tip
When evaluating an institutional knowledge claim, ask yourself: Does this institution have transparent methods? Is it subject to external review or accountability? Has it corrected its errors in the past? What incentives might shape its conclusions? These questions help you calibrate your trust — neither accepting everything uncritically nor rejecting everything reflexively. Trust should be proportional to evidence of reliability.
Epistemic Injustice
Epistemic injustice is a concept developed by the philosopher Miranda Fricker to describe situations in which someone is wronged specifically in their capacity as a knower. There are two primary forms:
-
Testimonial injustice occurs when a speaker's credibility is unfairly deflated due to prejudice. For example, when a patient's reports of pain are dismissed because of their race or gender, or when a young person's observations are disregarded simply because of their age.
-
Hermeneutical injustice occurs when someone lacks the shared concepts or language needed to make sense of their own experience. For example, before the concept of "sexual harassment" was named and defined, people who experienced it had difficulty articulating what was happening to them — not because they lacked intelligence, but because society lacked the conceptual resources.
Epistemic injustice damages everyone. When certain voices are systematically excluded, the resulting knowledge is impoverished. Medical research that excludes women from clinical trials produces incomplete knowledge. A historical narrative ignoring Indigenous perspectives is not just politically biased — it is epistemologically deficient.
Marginalized Knowledge
Marginalized knowledge refers to ways of knowing and knowledge traditions excluded from or devalued by dominant systems — Indigenous ecological knowledge, folk medicine, oral histories, and the experiential knowledge of oppressed communities.
This marginalization often results from historical power relations. Colonial powers suppressed Indigenous knowledge systems. Women's contributions to science were erased or misattributed. These exclusions were politically enforced, not epistemologically justified.
Recognizing marginalized knowledge does not mean accepting all claims uncritically. It means acknowledging that the criteria we use to evaluate knowledge — what counts as "evidence" or "rigor" — are themselves shaped by historical context. Epistemic justice requires examining whether those criteria systematically exclude valid forms of knowledge.
Representation
Representation in knowledge and power refers to who is present in the systems that produce, validate, and disseminate knowledge. When research teams, editorial boards, or algorithm design teams lack diversity, their outputs tend to reflect a narrow range of experiences.
This matters practically. Drug trials including only one ethnicity may not generalize. AI trained on one culture's values may be biased against others. Improving representation is not just social fairness — it is epistemic quality. More diverse knowledge-producing communities generate more robust and comprehensive knowledge.
Knowledge and Justice
Knowledge and justice are connected deeply. Access to knowledge is a prerequisite for meaningful democratic participation — citizens who lack accurate information cannot make informed decisions. Communities denied access to education are denied tools to advocate for their interests.
The production of knowledge is itself a site of justice or injustice. Whose questions get funded? Whose experiences get studied? Whose testimony gets believed? A just knowledge system would ensure not only equitable distribution but meaningful opportunities for all communities to participate in knowledge production and application.
Ethics of AI
The ethics of AI encompasses the moral questions raised by the design, deployment, and impact of artificial intelligence systems. Key ethical concerns include:
- Transparency: Should AI systems be required to explain their decisions? When an AI denies someone a loan or a job, does that person have a right to know why?
- Accountability: When an AI system causes harm — a self-driving car accident, a wrongful denial of medical treatment — who is responsible? The programmer? The company? The user?
- Consent: Do people have a right to know when they are interacting with an AI rather than a human? Do they have a right to opt out?
- Surveillance: How much data collection is justified in the name of security, efficiency, or knowledge production? Where is the line between useful data and invasive surveillance?
- Autonomy: As AI systems become more capable, how do we preserve meaningful human agency and decision-making? If an AI can write better essays, diagnose diseases more accurately, and make investment decisions more profitably than humans, what happens to human autonomy?
These questions lack simple answers. What matters epistemologically is that we ask them — refusing to treat technological development as inevitable and value-neutral, and instead subjecting it to the same critical scrutiny we apply to any knowledge claim.
Diagram: Ethical Dimensions of AI
Ethical Dimensions of AI
Type: diagram
sim-id: ai-ethics-dimensions
Library: p5.js
Status: Specified
Bloom Level: Evaluate (L5) Bloom Verb: Appraise Learning Objective: Appraise the ethical trade-offs involved in AI deployment by weighing competing values such as efficiency, transparency, fairness, and autonomy.
Instructional Rationale: A radar chart or multi-axis diagram allows students to see that AI ethics is not a single issue but a multi-dimensional space of competing values, where improving one dimension may create tension with another.
Visual elements: - A radar/spider chart with five axes: "Transparency," "Fairness," "Accountability," "Privacy," and "Autonomy" - Each axis scaled from 0 (low) to 10 (high) - Pre-loaded scenarios (e.g., "Medical Diagnosis AI," "Criminal Sentencing AI," "Social Media Algorithm") with different ethical profiles - A shaded polygon connecting the scores on each axis, making the ethical profile visually apparent
Interactive controls: - Dropdown to select different AI scenarios and see their ethical profiles - Sliders on each axis so students can design their own "ideal" ethical profile - A "Compare" button to overlay two scenarios for comparison - Text panel explaining the trade-offs for the selected scenario
Default state: "Social Media Algorithm" scenario loaded, showing its characteristic profile (high efficiency, low transparency, moderate fairness).
Color scheme: Axes in neutral gray, scenario polygons in translucent teal and amber for comparison, labels in dark text.
Responsive design: Canvas resizes to fit container width. Radar chart scales proportionally.
Implementation: p5.js with createSelect() for scenario selection and createSlider() for axis values
You've Got This!
The issues in this chapter can feel overwhelming — algorithms shaping what we know, power structures determining whose voices count, AI systems making decisions we cannot fully understand. It is completely natural to feel uncertain about how to navigate all of this. Remember: you do not need to have all the answers. What matters is that you are asking the right questions, examining assumptions, and refusing to take any knowledge system at face value. That is exactly what it means to think like an epistemologist.
Connecting Technology and Power: Who Benefits?
The two halves of this chapter — technology and power — are deeply intertwined. The most powerful technology companies control enormous quantities of data, deploy algorithms shaping the information environment of billions, and develop AI systems influencing consequential decisions. The question "Who controls the technology?" is inseparable from "Who controls the knowledge?"
- Data as power: Organizations that collect and control large datasets hold a form of power that is historically unprecedented. They know what you search for, what you buy, where you go, and whom you communicate with — and they can use that knowledge to predict and influence your behavior.
- Algorithmic gatekeeping: Social media algorithms determine what information billions of people see. These algorithms optimize for engagement — not for truth, not for fairness, not for democratic deliberation. The result is an information environment shaped by commercial incentives rather than epistemological values.
- The expertise gap: As AI systems become more complex, fewer people understand how they work. This creates an asymmetry of knowledge between those who design these systems and those who are affected by them — a form of power that is difficult to hold accountable.
| Intersection | How Technology Enables Power | How Power Shapes Technology |
|---|---|---|
| Data collection | Unprecedented surveillance capacity | Powerful actors decide what data to collect |
| Algorithmic curation | Shapes information access for billions | Commercial and political interests shape algorithm design |
| AI decision-making | Automates consequential decisions at scale | Those with resources determine AI development priorities |
| Digital literacy | Creates informed citizens who can resist manipulation | Educational inequalities limit access to digital literacy |
Digital Literacy: A Path Forward
In a world saturated with digital knowledge systems, digital literacy is an epistemological necessity — the ability to find, evaluate, create, and communicate information using digital technologies, with critical understanding of how those technologies work and whose interests they serve. It goes beyond basic computer skills to include:
- Source evaluation: Assessing the credibility, expertise, and potential biases of online sources
- Algorithmic awareness: Understanding that search results and social media feeds are curated, not neutral
- Data literacy: Interpreting data visualizations critically and recognizing when statistics are being misused
- Privacy awareness: Understanding what data is being collected about you and how it is being used
- AI literacy: Recognizing when you are interacting with AI-generated content and understanding its limitations
These competencies are essential for epistemic autonomy — the ability to form your own justified beliefs rather than having them shaped by systems you neither understand nor can evaluate.
Excellent Progress!
You have now explored one of the most complex and consequential intersections in all of epistemology — the place where technology, power, and knowledge meet. You understand how algorithms filter information, how data can encode bias, how power shapes whose knowledge counts, and why digital literacy matters more than ever. You are thinking like an epistemologist! In the next chapter, we will apply many of these ideas to the urgent challenge of misinformation and navigating the information age.
Chapter Summary
This chapter examined how digital technologies and power structures jointly shape knowledge production and access. Key takeaways include:
- Tools shape knowledge: Technologies are not neutral instruments — they influence what we can know and how we know it.
- Algorithms are gatekeepers: The invisible systems that curate our information environment make consequential judgments about relevance, credibility, and importance.
- Big data and AI create new epistemological challenges: Scale and automation can amplify existing biases while creating an illusion of objectivity.
- Power and knowledge are inseparable: Who controls resources, platforms, and institutions determines whose questions get asked and whose voices get heard.
- Epistemic injustice is real: Systematic exclusion from knowledge production harms both the excluded and the quality of knowledge itself.
- Digital literacy is essential: Critical engagement with digital systems is necessary for epistemic autonomy in the modern world.
Discussion Questions
- Think about the last time you searched for information online. How might the algorithm's choices about what to show you have shaped what you learned? What might you have missed?
- Can an AI system truly "know" something? Apply the criteria of justified true belief to an AI's outputs and evaluate whether they qualify as knowledge.
- Identify an example of marginalized knowledge from your own cultural context. What factors contributed to its marginalization, and what would be needed to recognize it as legitimate knowledge?
- Is it possible to design an algorithm that is truly unbiased? What would "unbiased" even mean in this context?
- How should societies balance the benefits of big data (better medical research, more efficient services) against the risks (surveillance, loss of privacy, algorithmic bias)?
- Think about an institution you trust and one you distrust. What factors explain the difference? Are your levels of trust epistemologically justified?