Skip to content
Open
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
48 commits
Select commit Hold shift + click to select a range
06cfffb
assignment1-diagram
Sdong23 Jan 26, 2025
fbb2ba8
assignment1
Sdong23 Jan 26, 2025
a594258
customer_purchase vs customer
Sdong23 Jan 27, 2025
351a690
make two lists and logical diagram
Sdong23 Jan 27, 2025
f1f309c
Write a query that returns everything in the customer table. */
Sdong23 Jan 27, 2025
4e2ffd7
Write a query that displays all of the columns and 10 rows from the c…
Sdong23 Jan 27, 2025
121b16c
Write a query that returns all customer purchases of product IDs 4 an…
Sdong23 Jan 27, 2025
b158f86
Write a query that returns all customer purchases and a new calculate…
Sdong23 Jan 27, 2025
bd35c73
write a query that outputs the product_id and product_name
Sdong23 Jan 27, 2025
8e13285
add a column to the previous query called pepper_flag that outputs a …
Sdong23 Jan 27, 2025
c7687d7
inner join
Sdong23 Jan 27, 2025
c6d57b9
count
Sdong23 Jan 27, 2025
6871742
generate list
Sdong23 Jan 27, 2025
918e7b1
make a temp
Sdong23 Jan 27, 2025
61d58d0
year and month
Sdong23 Jan 27, 2025
c447bca
2022April
Sdong23 Jan 27, 2025
6f8150f
ethics
Sdong23 Jan 27, 2025
ffa7195
diagram
Sdong23 Jan 27, 2025
6dbc772
Merge branch 'assignment-one' of https://github.com/Sdong23/sql into …
Sdong23 Jan 27, 2025
4a7dfb5
diagram init
Sdong23 Feb 2, 2025
18f0116
add employee, book, customer table
Sdong23 Feb 2, 2025
c84146b
update diagram and logic
Sdong23 Feb 2, 2025
066d0bc
Add shift table
Sdong23 Feb 2, 2025
fad7745
modify
Sdong23 Feb 2, 2025
cc0940d
section4
Sdong23 Feb 2, 2025
6f31ede
modified
Sdong23 Feb 2, 2025
eba383a
coalesce null and chenge unit
Sdong23 Feb 2, 2025
7cee35b
query for visit
Sdong23 Feb 2, 2025
616d58f
query for reverse visit
Sdong23 Feb 2, 2025
4c45062
most recent visit
Sdong23 Feb 2, 2025
357f1fe
dates for different product purchases
Sdong23 Feb 2, 2025
7b22086
Update Untitled Diagram.drawio
Sdong23 Feb 2, 2025
7a180b8
jar or organic
Sdong23 Feb 3, 2025
44c7be6
regexp number
Sdong23 Feb 3, 2025
85d07a5
best and worst day
Sdong23 Feb 3, 2025
6555ca2
product*5*customer
Sdong23 Feb 3, 2025
87a6525
product units
Sdong23 Feb 3, 2025
191daf5
insert
Sdong23 Feb 3, 2025
04ebf0f
delete older record
Sdong23 Feb 3, 2025
46a13a4
add product units
Sdong23 Feb 3, 2025
a863f70
add current_quantity
Sdong23 Feb 3, 2025
2a49d08
current quantity
Sdong23 Feb 3, 2025
104d048
restore
Sdong23 Feb 3, 2025
0fa21c9
Merge branch 'assignment-one' into assignment-two
Sdong23 Feb 3, 2025
4f513f7
modified for prompt 1
Sdong23 Feb 4, 2025
a8d1ccc
Update Untitled Diagram.drawio
Sdong23 Feb 4, 2025
f46943e
Added 2.drawio
Sdong23 Feb 4, 2025
3d0649c
shift table added
Sdong23 Feb 4, 2025
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
169 changes: 169 additions & 0 deletions 02_activities/assignments/2.drawio

Large diffs are not rendered by default.

16 changes: 15 additions & 1 deletion 02_activities/assignments/Assignment1.md
Original file line number Diff line number Diff line change
Expand Up @@ -205,5 +205,19 @@ Consider, for example, concepts of fariness, inequality, social structures, marg


```
Your thoughts...
In the article When Databases Get to Define Family by Rafia Qadri (2021), the author explores the implications of a digital database system implemented in Pakistan, which aims to track and define familial relationships through a centralized system. This system, known as the National Database and Registration Authority (NADRA), was designed to create a comprehensive digital record of families, providing better access to public services. However, as the article reveals, the technology inadvertently embeds certain value systems that have profound social implications. Through the lens of this case, we can examine how databases and data systems in our day-to-day lives often reflect underlying societal values, such as fairness, inequality, and social structures.

One of the most prominent value systems embedded in databases, like the one implemented by NADRA, is the idea of fairness. On the surface, the digitalization of family records aims to make government services more accessible and efficient. However, the concept of fairness is complicated when we consider who gets included in the system and who is marginalized.

In Pakistan, the NADRA database has, unintentionally, made it difficult for some families to fully benefit from state services. For example, families with limited access to technology, particularly in rural or underdeveloped regions, struggle to register their family members correctly in the digital system. Without proper documentation or understanding of how the database functions, marginalized groups—especially women, the illiterate, and economically disadvantaged—often find themselves excluded from benefits like healthcare, education, and social security. In this case, the system that was designed to promote fairness ends up exacerbating existing inequalities.

Databases like NADRA’s also reflect and reinforce certain social structures. In many societies, family structures are traditionally patriarchal, where men often hold the legal authority over family decisions. As the article illustrates, this digital system further solidifies these existing social hierarchies by using a male-centric model of family, where the father is often seen as the head of the household. This value system is embedded within the design of the database, influencing who gets recognized as a family’s primary representative.

The interaction of technology and society is crucial here. The database was created with the intention of bringing modernity and convenience to state processes. However, it doesn’t fully account for how deeply ingrained societal norms and structures may influence the way technology is used and experienced. By automating and digitalizing the process of recognizing family units, the system inadvertently disregards the complexity of familial relationships, particularly those that don’t fit neatly into the traditional mold of a nuclear family. For example, women who may be the primary caregivers or breadwinners of their families might not be properly acknowledged in the database system because the database relies on conventional norms about family leadership.

Another significant issue raised by Qadri is how databases can marginalize certain groups of people, especially in societies where access to technology and government services is unequal. Data systems, by their very nature, rely on the collection of information from individuals. Those without access to digital tools, like smartphones or reliable internet, are excluded from being able to register their information in the database. This digital divide creates a form of systemic marginalization. People from lower socioeconomic classes, women in conservative societies, or rural populations often find themselves excluded from digital platforms that are increasingly required for accessing basic services.

Furthermore, data systems can reinforce existing biases in how different groups are perceived and treated. In the case of NADRA, data systems may perpetuate discriminatory practices that disproportionately affect women or minority groups. For example, the system’s structure may not fully recognize non-traditional family structures, such as those involving blended families, single-parent households, or intergenerational households, thereby undermining their ability to receive state benefits. This structural bias built into the design of databases highlights the intersection of technology and societal values.

In conclusion, the values embedded in databases and data systems, such as fairness, inequality, and social structures, are not neutral. These systems are shaped by the societal norms and structures of the communities that create them. As illustrated by the NADRA system in Pakistan, what is intended to be a tool for inclusion can inadvertently perpetuate discrimination and marginalization, especially for those who are already disadvantaged. The article by Rafia Qadri emphasizes the need for greater awareness and sensitivity when developing digital systems to ensure that they do not inadvertently reinforce existing inequalities or exclude marginalized groups from their benefits. Ultimately, the intersection of technology and society requires a careful examination of how data systems shape our understanding of fairness, identity, and access to resources.
```
29 changes: 27 additions & 2 deletions 02_activities/assignments/Assignment2.md
Original file line number Diff line number Diff line change
Expand Up @@ -54,7 +54,9 @@ The store wants to keep customer addresses. Propose two architectures for the CU
**HINT:** search type 1 vs type 2 slowly changing dimensions.

```
Your answer...
Retain changes is type 2, overwrite is type 1.
type1: customer id, addresses
type2: customer id, current address, previous address, start_date for the address, end_date for the address.
```

***
Expand Down Expand Up @@ -182,5 +184,28 @@ Consider, for example, concepts of labour, bias, LLM proliferation, moderating c


```
Your thoughts...
The increasing integration of artificial intelligence (AI) and machine learning models into daily life presents a myriad of ethical challenges that often go unnoticed by the general public. In her article “Neural Nets Are Just People All the Way Down”, Vicki Boykis explores how neural networks, which are often hailed as autonomous or self-sustaining entities, are fundamentally reliant on human labor. She highlights critical ethical issues such as exploitation, bias, accountability, and the potential for harm in the proliferation of large language models (LLMs). This paper critically examines these concerns, illustrating how these issues intersect with societal inequalities, corporate responsibility, and the potential dangers posed by the widespread use of AI technologies.

One of the most pressing ethical issues discussed by Boykis is the exploitation of labor in the development of neural networks. While AI models such as neural nets may seem autonomous, they depend heavily on human workers to curate, label, and moderate the massive datasets that train these systems. Boykis argues that the general public tends to overlook the invisible labor that underpins the development of these technologies, which is often outsourced to workers in low-wage countries. Many of these workers are tasked with annotating images, moderating offensive content, or providing other services necessary for the training of AI systems.

This labor is not only underpaid but also emotionally taxing. Moderators, for example, often work with disturbing or harmful content, which can have significant psychological effects. Despite the enormous role these workers play in creating highly functional AI systems, they are often rendered invisible in discussions about the ethics of AI. The failure to recognize and properly compensate this labor speaks to a broader issue of labor exploitation in the tech industry, which raises significant ethical questions about the fairness and human cost of technological innovation.

Another key ethical concern in Boykis’ article revolves around the issue of bias in AI systems. Neural networks and large language models (LLMs) are trained on data collected from human sources. These datasets often reflect the biases, prejudices, and historical inequalities present in society. For instance, biased training data may result in algorithms that disproportionately disadvantage marginalized groups, reinforcing harmful stereotypes about race, gender, or socioeconomic status. Boykis argues that neural nets are not neutral; rather, they are shaped by the values and assumptions embedded in the data used to train them.

These biases can have real-world consequences, particularly in systems that make critical decisions about hiring, lending, policing, and healthcare. A hiring algorithm, for example, may prefer male candidates over female candidates, or a loan application algorithm may discriminate against people of color based on biased historical data. These issues underscore the ethical imperative for developers to actively address and mitigate bias in AI systems to ensure they do not perpetuate or exacerbate existing societal inequalities.

As AI systems, particularly large language models (LLMs), become more ubiquitous, there are growing concerns about the potential risks they pose in terms of content moderation and the spread of misinformation. Boykis notes that these models, while capable of generating impressively coherent and human-like text, can also be misused to spread disinformation, create fake news, and impersonate individuals. The ability of LLMs to generate persuasive text makes them powerful tools for malicious actors seeking to manipulate public opinion or deceive the public.

The ethical implications of LLM proliferation extend beyond just the creation of false content; they also raise questions about the control of information. Who decides what information is disseminated by AI systems? Companies that develop these technologies, such as OpenAI, Google, and others, hold significant power over what content is generated and shared. Given the reach of AI-generated content, this concentration of power raises concerns about the potential censorship or biased dissemination of information, where certain voices and perspectives may be suppressed or amplified based on corporate interests.

A critical ethical issue in the deployment of neural networks and LLMs is the question of accountability. Boykis draws attention to the challenges associated with determining responsibility when AI systems make harmful or biased decisions. If an AI system recommends the wrong candidate for a job or denies a loan to an individual based on biased data, who is responsible? The developers of the system? The corporations deploying it? Or the AI itself? The lack of transparency in many AI systems makes it difficult to pinpoint the source of the problem, complicating efforts to hold anyone accountable.

The question of accountability becomes even more pressing as AI systems gain more autonomy and decision-making capabilities. If these systems are capable of making high-stakes decisions without human intervention, it becomes ethically problematic to absolve developers and companies of responsibility when things go wrong. Moreover, the lack of explainability in many neural networks — often described as the “black box” problem — makes it difficult for stakeholders to understand how these models arrive at their conclusions. This lack of transparency exacerbates concerns about accountability and calls for greater regulation and oversight of AI technologies.

The Intersection of Technology and Society
The ethical issues surrounding neural networks and AI are not only technological but also deeply societal. Boykis underscores the political nature of AI, highlighting how these technologies reflect the interests and values of the people and corporations that design them. The deployment of AI systems can have far-reaching consequences for societal structures, particularly in areas such as education, employment, and healthcare. The question of who benefits from AI advancements — and who is left behind — is critical to understanding the ethical ramifications of these technologies.

As AI systems become more embedded in society, they risk entrenching existing inequalities. For example, people without access to advanced AI tools may find themselves marginalized in an increasingly automated world. Moreover, AI systems that prioritize corporate profit over the public good may exacerbate social inequalities, raising important ethical concerns about the intersection of technology, capitalism, and social justice.

The ethical issues surrounding neural networks and large language models are complex and multifaceted. From labor exploitation and data bias to the risks of misinformation and accountability, these technologies present significant challenges that require careful consideration. As AI continues to evolve, it is crucial that developers, policymakers, and society at large address these ethical issues, ensuring that AI systems are developed and deployed in ways that benefit all people, not just those who control the technology. Only by confronting these challenges head-on can we create a future where AI serves the public good rather than exacerbating existing inequalities and injustices.
```
Loading