How could #ArtificialIntelligence best be regulated, besides with copyright and creative commons law for data, & to protect the public domain and creators?
— WorldUnivandSch (@WorldUnivAndSch) May 26, 2023
How ~200 online CC-4 licensed #MITOCW-centric wiki @WorldUnivAndSchhttps://t.co/HxWJoXw3Bj s' faculties could teach this?
I asked #GoogleBard https://bard.google.com/ & #ChatGPT how #ArtificialIntelligence could best be regulated, besides with copyright law @lessig. Interesting: https://scott-macleod.
I asked #GoogleBardhttps://t.co/NRQv78FPGH& #ChatGPT how #ArtificialIntelligence could best be regulated, besides with copyright law @lessig. Interesting: https://t.co/7c6aqyWofv& with ~200 online #CC4licensed#MITOCW-centric wiki @WorldUnivAndSchhttps://t.co/HxWJoXw3Bj s
— WorldUnivandSch (@WorldUnivAndSch) May 26, 2023
https://twitter.com/
https://twitter.com/
https://twitter.com/
https://twitter.com/
https://twitter.com/WUaSPress/
I asked #GoogleBardhttps://t.co/mihcn1moiB& #ChatGPT how #ArtificialIntelligence could best be regulated, besides with copyright law @lessig. Interesting: https://t.co/tssaQhXXXI& with ~200 online #CC4licensed#MITOCW-centric wiki @WorldUnivAndSchhttps://t.co/d3Yst4HCDT s
— the Open Band (Berkeley) (@TheOpenBand) May 26, 2023
- Scott GK MacLeod
* *
Google Bard & Chat GPT - How best could responsible artificial intelligence be regulated, besides with copyright and creative commons law for data, and to protect the public domain and creators?
There are a number of ways that responsible artificial intelligence could be regulated, besides with copyright and creative commons law for data, and to protect the public domain and creators. These include:
- Transparency: AI systems should be transparent in their operation, so that users can understand how they work and make informed decisions about their use. This could be achieved through requirements for clear and concise documentation, as well as the ability for users to inspect the data and algorithms used by the system.
- Accountability: There should be clear accountability for the development and use of AI systems. This could be achieved through requirements for developers to identify themselves and their affiliations, as well as the ability for users to hold developers accountable for any harm caused by their systems.
- Fairness: AI systems should be designed to be fair and unbiased, and should not discriminate against any particular group of people. This could be achieved through requirements for systems to be tested for bias, and for developers to take steps to mitigate any bias that is found.
- Privacy: AI systems should respect the privacy of users, and should not collect or use personal data without consent. This could be achieved through requirements for systems to have clear privacy policies, and for developers to take steps to protect user data.
- Security: AI systems should be secure, and should be protected from hacking and other forms of attack. This could be achieved through requirements for systems to be designed with security in mind, and for developers to take steps to patch any security vulnerabilities that are found.
It is important to note that these are just a few of the ways that responsible AI could be regulated. The specific requirements that are needed will vary depending on the specific application of AI. However, by taking steps to ensure transparency, accountability, fairness, privacy, and security, we can help to ensure that AI is used for good and not for harm.
In addition to the above, there are a number of other factors that should be considered when regulating responsible AI. These include:
- The potential for AI to be used for harmful purposes: AI has the potential to be used for a wide range of harmful purposes, including discrimination, surveillance, and warfare. It is important to consider these potential harms when developing regulations for AI.
- The need to balance innovation with safety: AI is a rapidly evolving field, and it is important to balance the need to encourage innovation with the need to ensure safety. Regulations should be designed in a way that does not stifle innovation, but that does help to ensure that AI is used safely.
- The need to be adaptable to the changing nature of AI: AI is a rapidly changing field, and it is important to ensure that regulations are adaptable to the changing nature of AI. Regulations should be designed in a way that allows them to be updated as AI technology evolves.
Regulating responsible artificial intelligence (AI) involves addressing a range of ethical, legal, and societal considerations. While copyright and creative commons laws can protect data ownership and incentivize innovation, they might not be sufficient to fully address the complexities associated with regulating AI. Here are some additional approaches that can be considered:
Ethical Frameworks: Develop and promote ethical guidelines for AI development and deployment. These frameworks can outline principles such as fairness, transparency, accountability, privacy, and human-centric design. They serve as a foundation for responsible AI practices and can be adopted voluntarily or enforced through regulation.
Risk Assessments and Impact Statements: Mandate organizations developing or using AI to conduct thorough risk assessments and impact statements. These assessments should evaluate potential biases, discrimination, safety risks, and societal impacts. Regulatory bodies can review and verify these assessments to ensure responsible AI practices.
Transparency and Explainability: Enforce regulations that require AI systems to be transparent and explainable. This means organizations should provide clear explanations of how AI systems make decisions and allow for external audits or third-party scrutiny. It enables users and regulators to understand and challenge the outcomes produced by AI algorithms.
Data Privacy and Security: Strengthen data protection regulations to safeguard individuals' privacy rights. This includes ensuring informed consent for data collection and usage, implementing robust data security measures, and holding organizations accountable for breaches or misuse of personal data.
Algorithmic Audits and Certification: Establish mechanisms to conduct audits of AI algorithms and systems. Independent auditors can assess the fairness, accuracy, and bias within AI systems. Certification processes can provide a seal of approval for responsible AI systems, encouraging compliance with ethical standards.
Governance and Oversight: Create regulatory bodies or agencies specifically dedicated to overseeing AI development and deployment. These bodies can monitor compliance, set standards, and investigate incidents or violations. They should involve interdisciplinary experts, including AI researchers, ethicists, legal professionals, and representatives from civil society.
Public Engagement and Education: Foster public awareness and understanding of AI technologies through education and public engagement initiatives. Promote discussions on AI's potential benefits, risks, and ethical implications, ensuring that public concerns and values are considered in regulatory decision-making processes.
International Collaboration: Encourage collaboration and harmonization of AI regulations at the international level. AI development and deployment are global issues, and cross-border coordination can help establish consistent standards and prevent regulatory loopholes.
It's important to note that regulation should be dynamic and adaptive to keep pace with evolving AI technologies and their potential impacts. Balancing innovation with responsible AI practices requires ongoing collaboration between policymakers, industry stakeholders, researchers, and the public.
Creative Commons (CC) licenses are a set of copyright licenses that allow creators to share their work with others while still retaining some control over how it is used. CC licenses are designed to be flexible and adaptable, and they can be used to protect a wide range of creative works, including text, music, images, and videos.
CC licenses can be used to regulate AI in a number of ways. For example, CC licenses can be used to ensure that AI systems are not used to create content that infringes on the copyrights of others. Additionally, CC licenses can be used to ensure that AI systems are not used to collect or use personal data without consent.
However, CC licenses are not a silver bullet for regulating AI. For example, CC licenses cannot prevent AI systems from being used to create harmful content, such as hate speech or propaganda. Additionally, CC licenses can be difficult to enforce, especially in cases where AI systems are used to create content that is not easily identifiable as infringing.
Overall, CC licenses can be an effective tool for regulating AI in some cases. However, it is important to note that CC licenses are not a perfect solution, and they should be used in conjunction with other regulatory measures to ensure that AI is used responsibly.
Here are some of the challenges of regulating AI using CC licenses:
- It can be difficult to identify the copyright holder of the data used to train an AI system. This is because data is often collected from a variety of sources, and it may not be clear who owns the copyright to each piece of data.
- It can be difficult to enforce CC licensing requirements. This is because AI systems can be used to create content that is difficult to identify as infringing. Additionally, it may be difficult to track down the users of AI systems who are infringing on copyrights.
- CC licenses can stifle innovation. This is because CC licenses can make it more expensive and difficult to develop and use AI systems.
Despite these challenges, CC licenses can be an effective tool for regulating AI in some cases. By carefully considering the challenges and benefits of CC licenses, policymakers can develop effective regulations that help to ensure that AI is used responsibly.
While Creative Commons licensing can play a role in regulating the use and sharing of AI data, it alone may not be sufficient to effectively regulate artificial intelligence as a whole. Creative Commons licenses are primarily designed to govern the copyright and distribution of creative works, including data. They provide a framework for granting permissions and specifying the terms of use for data or content.
However, regulating AI involves addressing a broader range of considerations beyond data licensing. Here are some reasons why Creative Commons licensing may have limitations in effectively regulating AI:
Scope: AI regulation encompasses various aspects beyond data, such as algorithmic transparency, fairness, accountability, safety, privacy, and bias mitigation. Creative Commons licenses primarily focus on copyright and licensing issues and may not directly address these broader concerns.
Technical Complexity: AI systems are highly complex and often involve numerous components, including data, algorithms, models, and infrastructure. Regulating AI requires a nuanced understanding of these technical aspects, which goes beyond the scope of Creative Commons licensing.
Dynamic Nature: AI technologies are rapidly evolving, and Creative Commons licenses are typically static agreements. Regulating AI requires adaptive frameworks that can keep pace with emerging challenges and new advancements in technology.
Legal Frameworks: AI regulation may require specific legal frameworks that address liability, accountability, safety standards, and governance mechanisms. While Creative Commons licenses offer permissions and conditions for data use, they may not provide the necessary legal structures to address these broader regulatory aspects.
To effectively regulate AI, it is important to consider a combination of approaches, including ethical frameworks, transparency requirements, data protection laws, auditing mechanisms, and oversight bodies. These approaches aim to ensure responsible AI development and deployment while addressing the societal, ethical, and legal challenges associated with AI technologies.
Copyright licensing of AI data can be an effective way to regulate AI in some cases. For example, copyright licensing can be used to ensure that AI systems are not used to create content that infringes on the copyrights of others. Additionally, copyright licensing can be used to ensure that AI systems are not used to collect or use personal data without consent. However, copyright licensing is not a silver bullet for regulating AI. For example, copyright licensing cannot prevent AI systems from being used to create harmful content, such as hate speech or propaganda. Additionally, copyright licensing can be difficult to enforce, especially in cases where AI systems are used to create content that is not easily identifiable as infringing.
Overall, copyright licensing can be an effective tool for regulating AI in some cases. However, it is important to note that copyright licensing is not a perfect solution, and it should be used in conjunction with other regulatory measures to ensure that AI is used responsibly.
Here are some of the challenges of regulating AI using copyright licensing:
- It can be difficult to identify the copyright holder of the data used to train an AI system. This is because data is often collected from a variety of sources, and it may not be clear who owns the copyright to each piece of data.
- It can be difficult to enforce copyright licensing requirements. This is because AI systems can be used to create content that is difficult to identify as infringing. Additionally, it may be difficult to track down the users of AI systems who are infringing on copyrights.
- Copyright licensing can stifle innovation. This is because copyright licensing can make it more expensive and difficult to develop and use AI systems.
Despite these challenges, copyright licensing can be an effective tool for regulating AI in some cases. By carefully considering the challenges and benefits of copyright licensing, policymakers can develop effective regulations that help to ensure that AI is used responsibly.
Regulating artificial intelligence (AI) solely through copyright licensing of its data may not be sufficient or effective in addressing the complexities of AI regulation. While copyright can provide legal protection for original works, including data, it primarily focuses on issues of ownership, usage rights, and distribution. Here are some reasons why relying solely on copyright licensing of AI data may have limitations:
Limited Scope: AI regulation involves considerations beyond data ownership and licensing. It encompasses a wide range of ethical, legal, and societal aspects, including algorithmic transparency, fairness, accountability, safety, privacy, bias mitigation, and more. Copyright licensing alone does not address these broader concerns.
Data Independence: AI systems often rely on diverse datasets from various sources, including public domain data, open data, and proprietary data. While copyright licensing may be applicable to proprietary data, it may not cover data from other sources, limiting its effectiveness as a regulatory mechanism for AI as a whole.
Accountability and Transparency: AI systems are often complex and involve not just data but also algorithms, models, and decision-making processes. Regulation needs to go beyond data ownership to ensure accountability, transparency, and explainability in AI systems. Copyright licensing does not inherently address these aspects.
Dynamic Nature: AI technologies are rapidly evolving, and copyright licenses are typically static agreements. AI regulation requires flexible and adaptable frameworks to keep pace with technological advancements, emerging challenges, and evolving societal needs.
Interdisciplinary Approach: Regulating AI effectively requires the involvement of various stakeholders, including researchers, ethicists, legal experts, policymakers, and the public. Copyright licensing, while important for data usage, may not provide the necessary mechanisms for interdisciplinary collaboration and comprehensive AI regulation.
To effectively regulate AI, it is important to consider a holistic approach that encompasses a range of strategies and measures. This can include the development of ethical frameworks, transparency requirements, data protection laws, algorithmic audits, governance mechanisms, and public engagement initiatives. These approaches address the multifaceted challenges associated with AI technologies and promote responsible and beneficial AI deployment.
Artificial intelligence (AI) is a branch of computer science that deals with the creation of intelligent agents, which are systems that can reason, learn, and act autonomously. Machine learning (ML) is a subset of AI that deals with the development of algorithms that can learn from and make predictions on data.
In simple terms, AI is the ability of a machine to learn and perform tasks that are typically associated with human intelligence, such as reasoning, planning, and learning. Machine learning is a subset of AI that uses algorithms to learn from data and make predictions.
Machine learning algorithms are trained on data sets that contain examples of the desired output. For example, a machine learning algorithm that is trained to classify images of cats and dogs would be trained on a data set that contains images of cats and dogs, along with labels that indicate whether each image is a cat or a dog. Once the algorithm is trained, it can be used to classify new images of cats and dogs.
Machine learning is a powerful tool that can be used to solve a wide variety of problems. Some of the most common applications of machine learning include:
- Predictive analytics: Machine learning can be used to predict future events, such as customer behavior, product demand, and fraud.
- Natural language processing: Machine learning can be used to understand and process human language, such as text and speech.
- Computer vision: Machine learning can be used to identify objects and scenes in images and videos.
- Robotics: Machine learning can be used to control robots and other autonomous systems.
Machine learning is a rapidly growing field with new applications being developed all the time. As the amount of data available continues to grow, machine learning is becoming increasingly powerful and capable.
Here are some examples of AI and machine learning in action:
- Self-driving cars: Self-driving cars use a variety of AI and machine learning technologies to navigate the road and avoid obstacles.
- Virtual assistants: Virtual assistants like Siri and Alexa use machine learning to understand your voice commands and respond accordingly.
- Fraud detection: Machine learning is used to detect fraudulent activity in financial transactions.
- Medical diagnosis: Machine learning is used to help doctors diagnose diseases more accurately.
- Product recommendations: Machine learning is used to recommend products to customers based on their past purchases and browsing history.
These are just a few examples of the many ways that AI and machine learning are being used today. As these technologies continue to develop, we can expect to see even more innovative applications in the future.
Artificial intelligence (AI) and machine learning (ML) are related concepts that are often used interchangeably but have distinct meanings. Here's an overview of both:
Artificial Intelligence (AI): AI refers to the development of computer systems or machines that can perform tasks that typically require human intelligence. It involves creating intelligent machines capable of perceiving, reasoning, learning, problem-solving, and making decisions. AI can be classified into two types:
Narrow AI or Weak AI: This type of AI is designed to perform specific tasks or functions within a limited domain. Examples include voice assistants like Siri or Alexa, recommendation systems, and image recognition algorithms.
General AI or Strong AI: General AI aims to possess human-level intelligence and the ability to understand, learn, and apply knowledge across different domains. This level of AI remains largely hypothetical and does not currently exist in practical applications.
Machine Learning (ML): Machine learning is a subset of AI that focuses on enabling machines or computer systems to learn from data and improve their performance without being explicitly programmed. Instead of being explicitly programmed with specific instructions, ML algorithms use statistical techniques to automatically learn patterns and make predictions or decisions based on the data they receive.
ML algorithms can be broadly categorized into three types:
Supervised Learning: In this approach, the ML model learns from labeled training data, where each data point is associated with a known target or output. The model learns to map input features to the corresponding outputs and can make predictions on unseen data.
Unsupervised Learning: Unsupervised learning involves training ML models on unlabeled data, where the model learns to identify patterns, group similar data, or discover hidden structures without explicit guidance. Clustering and dimensionality reduction are common unsupervised learning techniques.
Reinforcement Learning: Reinforcement learning involves training an agent to interact with an environment and learn optimal actions through a system of rewards and punishments. The agent learns to maximize its rewards by trial and error and reinforcement from the environment.
Machine learning algorithms play a crucial role in many AI applications, enabling systems to learn and improve their performance over time. They are used in various domains, including image and speech recognition, natural language processing, autonomous vehicles, recommendation systems, and many more.
It's important to note that while ML is a key component of AI, AI encompasses a broader range of techniques and approaches beyond just machine learning.
- Scott GK MacLeod
*
Scott MacLeod <sgkmacleod@worlduniversityandschool.org>Thu, May 25, 2023 at 10:01 PM
To: "Lessig, Lawrence"<lessig@law.harvard.edu>
Cc: Allison Ryder <amryder@mit.edu>, Giesela Rühl <giesela.ruehl@hu-berlin.de>, "Mills.Steven@bcgfed.com"<mills.steven@bcgfed.com>, Lydia Pintscher <lydia.pintscher@wikimedia.de>, Roland Vogl <rvogl@law.stanford.edu>
I asked Google Bard and Chat GPT related questions such as the following -
"How best could responsible artificial intelligence be regulated, besides with copyright and creative commons law for data, and to protect the public domain and creators?"
Am curious how ~200 online CC-4 licensed MIT OCW-centric (wiki) World Univ & Sch Law Schools' facilities could teach iteratively regulating artificial intelligence - and to protect both the public domain and creators.
- Scott GK MacLeod
Founder, President, CEO & Professor
CC-4 licensed MIT OCW-centric. Wiki,
World University & School (WUaS)
- PO Box 442, Canyon, CA 94516
- 210 East End Avenue, Pittsburgh, PA 15221
1) non-profit World University and School - http://worlduniversityandschool.org
- https://twitter.com/WorldUnivAndSch
2) for profit general stock company WUaS Corporation in CA - http://worlduniversityandschool.org/AcademicPress.html
- https://twitter.com/WUaSPress
(o) 415 480 4577 - sgkmacleod@worlduniversiryandschool.org
(m) 412 478 0116 - sgkmacleod@gmail.com
Law Schools at World University and School (planned in main languages in them)
Afghanistan Law School at WUaS: https://wiki.worlduniversityandschool.org/wiki/Afghanistan_Law_School_at_WUaS
Brazil Law School at WUaS: https://wiki.worlduniversityandschool.org/wiki/Brazil_Law_School_at_WUaS
China Law School at WUaS: https://wiki.worlduniversityandschool.org/wiki/China_Law_School_at_WUaS
Egypt Law School at WUaS: https://wiki.worlduniversityandschool.org/wiki/Egypt_Law_School_at_WUaS
India Law School at WUaS: https://wiki.worlduniversityandschool.org/wiki/India_Law_School_at_WUaS
Jamaica Law School at WUaS: https://wiki.worlduniversityandschool.org/wiki/Jamaica_Law_School_at_WUaS
Mexico Law School at WUaS: https://wiki.worlduniversityandschool.org/wiki/Mexico_Law_School_at_WUaS
World University Law School: https://wiki.worlduniversityandschool.org/wiki/World_University_Law_School
https://worlduniversityandschool.blogspot.com/search/label/Law%20Schools
https://scott-macleod.blogspot.com/search/label/Law%20Schools
On Thu, May 25, 2023 at 5:28 AM Lessig, Lawrence <lessig@law.harvard.edu> wrote:
I am far from policy making with Creative Commons these days, but my own view is that these important issues should not be regulated by copyright licenses. We should be pushing for an appropriate framework of protection — both for the public domain and creators — outside of the frame of licenses grounded in the act of copying.
______________________________________
Lawrence Lessig
Roy L. Furman Professor of Law and Leadership | Harvard Law School
1563 Massachusetts Avenue │ Cambridge, MA 02138
(617) 496-8853 │ (617) 496-5156 (fax) │ @lessig
On May 23, 2023 at 11:11:59, Scott MacLeod (sgkmacleod@worlduniversityandschool.org) wrote:
Dear Allison, Steve, Professor Larry Lessig, (Giesela, Lydia, Roland),
Thanks for your email, Allison, and as the MIT Sloan School of Management director of the Big Ideas' project. I'll seek to see your presentations again when they are posted - https://www.youtube.com/playlist?list=PL48E61C121CAD0E1B. Regarding my question about Google artificial intelligence and your 'Responsible AI' (RAI) methodologies, and your "MIT SMR is creating its own guidelines for the use of generative AI, though my gut-reaction answer to your question about using it in our work is that our content is often sensitive and as we own the copyright, releasing it for use by LLMs ... ," I find Google's TensorFlow Responsible AI 4 principles (below too in the PPS) - fairness, interpretability, privacy and security (https://www.tensorflow.org/responsible_ai) - germane, for example, to MIT OCW-centric wiki World Univ & Sch's growth of #WUaSArtificialIntelligence in 200 countries and in their main languages, and eventually in all 7151 known living languages, and e.g. as emerging AI & ML Large Language Models.
Professor Lessig, Allison Ryder, all, how do you think Creative Commons' licensing and law will inform the developing fields of responsible AI, and ethical AI, and related policy and regulatory frameworks, not only in the USA, but also in the European Union, and in all 200 countries, and as founder of Creative Commons' licensing / law? CC-4 MIT OpenCourseWare licensing, in 5 languages presently -
Translated Courses
Spanish
Traditional Chinese
Turkish
Korean
http://ocw.abu.edu.ng/courses/translated-courses/
... allows for 1) sharing (freely with attribution), 2) adapting (eg using it as data for AI), but 3) non-commercially, for example (so MIT OCW-centric wiki World Univ & Sch can't seek tuition from matriculating students), and CC-0 licensed Wikidata, as Wikipedia's backend structured knowledge database with querying, in its 300 languages, allows for artificial intelligence and machine learning data usage, in a related new responsible AI or ethical AI way to CC-4 licensing (https://creativecommons.org/about/cclicenses/ and see regarding CC-4 licensing https://ocw.mit.edu and its https://creativecommons.org/licenses/by-nc-sa/4.0/). How might this inform "MIT SMR is creating its own guidelines for the use of generative AI" as they emerge, Allison? Has or will the MIT Sloan Management Review looked yet at Creative Commons' licensings' impact on responsible AI ... and perhaps by way of comparison with Google's approaches to responsible AI, as a mostly non-creative commons' organization (although the CC-4 licensed "CS First with Google at World Univ & Sch" - https://csfirst.withgoogle.com/s/en/home - Scratch programming language course is probably engaging Google's TensorFlowAI or similar)?
Regarding, Allison, "... as far as who is considered an RAI “Leader”," am appreciating the title of your MIT SMR "To Be a Responsible AI Leader, Focus on Being Responsible," and am also appreciating Google very much in these regards, and regarding my questions in your great presentation. Am seeking for MIT OCW-centric wiki World Univ & Sch to become a Responsible AI Leader, and regarding the categories that Steve Mills' shared especially.
Steve, regarding your email address at the Boston Consulting Group - Mills.Steven@bcgfed.com - are you consulting too for the Boston Federal Reserve?
In these regards, and in MIT OCW-centric wiki World University and School seeking to code for all 7.9 billion people on the planet, each a Wikidata PIN #, in all 200 countries and among speakers of all 7151 known living languages, and even to distribute a main single cryptocurrency, like Stanford Mine Pi cryptocurrency - and especially to end poverty worldwide via UBI experiments for example (and to facilitate wiki people-to-people free universal education among speakers of all 7151 known living languages) - what role would the MIT SMR guidelines play in examining such new approaches to these WUaS responsible AI developments?
Thank you for your MIT SMR Big Ideas' project, Allison, and thank you for your Stanford Law CodeX presentations, Allison, Giesela and Steve, especially.
Best wishes, Scott
https://scott-macleod.blogspot.com/2023/05/goldenrod-gigantea-subsp-serotina-ne.html
- Scott GK MacLeod
Founder, President, CEO & Professor
CC-4 licensed MIT OCW-centric. Wiki,
World University & School (WUaS)
- PO Box 442, Canyon, CA 94516
- 210 East End Avenue, Pittsburgh, PA 15221
1) non-profit World University and School - http://worlduniversityandschool.org
2) for profit general stock company WUaS Corporation in CA - http://worlduniversityandschool.org/AcademicPress.html
PS
Stanford Mine Pi cryptocurrency, and for 7.9 billion people as responsible AI especially -
Steve, especially, in related vein, and regarding growing World Univ & Sch, and WUaS Monthly Business Meeting on Saturday, May 18,2023, could info@worlduniversityandschool.org email account workspace becoming disabled after I sent out an agenda for MBM Thursday be because a http://worlduniversityandschool.org/web/ WordPress site has been spamming info@worlduniversityandschool.org (which I've been locked out of so can't correct this) and after I've inquired with Peter Norvig, Lydia Pintscher, Ed Smyth MD (chief medical officer at MIT OCW-centric wiki World Univ & Sch), & MIT DCI director Neha Narula others months ago about turning this spewing WUaS database web site address into a way to code for all 7.9 billion people and to end poverty and create Wiki free universal education at WUaS among speakers of all 7151 living languages?
Could a whole group of coders, from a variety of different organizations (Google Education regarding this email account, Wikidata/Wikimedia, Stanford Mine Pi, various US federal reserve banks?), be creating IT infrastructure for WUaS Stanford Mine Pi to become a main near universal cryptocurrency in 200 countries from here - https://wiki.worlduniversityandschool.org/wiki/You_at_World_University and the above related WUaS WordPress site and database?
WITH -
Stanford Mine Pi cryptocurrency databases, with which WUaS is partnering, and see -
"I am sending you 1π! Pi is a new digital currency developed by Stanford PhDs, with over 35 million members worldwide. To claim your Pi, follow this link https://minepi.com/sgkmac and use my username (sgkmac) as your invitation code."
(And possibly for molecules and cells, etc., in a realistic virtual earth for genetics -
https://scott-macleod.blogspot.com/2023/05/violet-nj-new-jersey-state-flower.html ?)
PPS
TensorFlow - Responsible AI
https://www.tensorflow.org/responsible_ai
TensorFlow
Thanks for tuning in to Google I/O.
Learn how to integrate Responsible AI practices into your ML workflow using TensorFlow
TensorFlow is committed to helping make progress in the responsible development of AI by sharing a collection of resources and tools with the ML community.
What is Responsible AI?
The development of AI is creating new opportunities to solve challenging, real-world problems. It is also raising new questions about the best way to build AI systems that benefit everyone.
Recommended best practices for AI
Designing AI systems should follow software development best practices while taking a human-centered
approach to ML
Fairness
As the impact of AI increases across sectors and societies, it is critical to work towards systems that are fair and inclusive to everyone
Interpretability
Understanding and trusting AI systems is important to ensuring they are working as intended
Privacy
Training models off of sensitive data needs privacy preserving safeguards
Security
Identifying potential threats can help keep AI systems safe and secure
https://www.tensorflow.org/responsible_ai
PPPS
"This makes Wikidata important to the internet..."
PPPPS
Lawrence Lessig, Founder / Board Member Emeritus, Creative Commons
https://creativecommons.org/author/lessig/
PPPPPS
"To Be a Responsible AI Leader, Focus on Being Responsible"
https://sloanreview.mit.edu/projects/to-be-a-responsible-ai-leader-focus-on-being-responsible/
PPPPPPS
"Machine learning, explained"
https://mitsloan.mit.edu/ideas-made-to-matter/machine-learning-explained
Thanks, Scott
with much of these communications here -
https://scott-macleod.blogspot.com/2023/05/goldenrod-gigantea-subsp-serotina-ne.html
On Mon, May 22, 2023 at 2:02 PM Allison Ryder <amryder@mit.edu> wrote:
*
*
https://en.wikipedia.org/wiki/Rosa_blanda
https://commons.wikimedia.org/wiki/Category:Rosa_blanda
https://en.wikipedia.org/wiki/Rosa_arkansana
https://commons.wikimedia.org/wiki/Rosa_arkansana
https://commons.wikimedia.org/wiki/Category:Rosa_arkansana
....