265SmithWatt 75Neumann 55.YunusAbed , AI20s.com JHDHFL 20
KingCharlesLLM DeepLearning009 NormanMacrae.net EconomistDiary.com Abedmooc.com
source
, Chief Executive Officer, Lablup
, Chief Technology Officer, Lablup, Inc. at nvidia gtc 2024 march
LLM descriptions by GEMINI
Falcon-180B, a large language model (LLM) developed by the Technology Innovation Institute (TII) in Abu Dhabi. It's the largest model within the Falcon LLM family.
Here's a breakdown of key points about Falcon-180B:
Size and Capabilities: With 180 billion parameters, it's one of the largest publicly available LLMs, potentially offering superior performance in tasks like text generation, translation, question answering, and code generation compared to smaller models.
Open-Source: Like other Falcon LLMs, Falcon-180B is open-source, allowing researchers and developers to access and modify the underlying code, fostering innovation in the field of AI.
Accessibility: The open-source nature removes licensing fees, making it potentially more accessible for non-commercial users compared to some closed-source LLMs.
Challenges: Running and training large models like Falcon-180B requires significant computing resources. Specialized hardware like GPUs might be needed, and training can be expensive.
Competition: While large, Falcon-180B faces competition from other large open-source models like Megatron-Turing NLG from Google AI, and closed-source models like GPT-4 from OpenAI.
Overall:
Falcon-180B is a powerful open-source LLM with the potential to advance various AI applications. However, its large size presents challenges in terms of computational resources required for training and running the model.
Here are some additional points to consider:
Fine-Tuning: Falcon-180B is a pre-trained model, but it can be further fine-tuned for specific tasks using specialized datasets.
Community and Support: The open-source nature allows for a larger developer community to contribute to the model's improvement and create tools for easier use.
Ethical Considerations: As with all large language models, ethical considerations like potential bias and misuse need to be addressed when using Falcon-180B.
I hope this information is helpful! Let me know if you have any other questions about Falcon-180B or large language models in general
here isn't a single, completely open-source Western LLM that directly compares to the size and capabilities of Falcon-180B (180 billion parameters) at this time (April 2024). Here's a breakdown of the landscape:
Western Open-Source LLMs:
Reasons for Size Discrepancies:
Alternatives and Considerations:
Overall:
The landscape of open-source LLMs is constantly evolving. While the West might not have a completely open-source model matching Falcon-180B's size yet, there are significant contributions and alternative options available.
Tags:
Views: 20
Llama2 : Meta is a family of large language models (LLMs) developed by Meta (formerly Facebook) AI. It's a significant player in the open-source LLM scene, offering capabilities similar to other well-known models like GPT-3 and PaLM 2. Here's a breakdown of key points about Llama2:
Open-Source and Accessibility:
Multiple Model Sizes:
Focus on Safety and General-Purpose Tasks:
Chat-Optimized Variant - Llama2-chat:
Comparison to Other LLMs:
Availability:
Overall:
Llama2 is a promising open-source LLM with a focus on accessibility, safety, and general-purpose capabilities. Its various sizes and chat-optimized variant cater to different needs. While it might not be the single most powerful LLM, its open-source nature and ongoing development make it a valuable addition to the AI landscape.
Here are some resources for further exploration:
Do you have any other questions about Llama2 or large language models in general?
Claude2 Anthropic
Claude2 is a large language model (LLM) developed by Anthropic, a research company founded by former OpenAI researchers. Here's a breakdown of its key features and potential:
Improved Performance:
Longer Responses and Stronger Memory:
Focus on Safety and Explainability:
Limited Public Availability:
Competition and Comparisons:
Uncertainties and Ongoing Development:
Ethical Considerations:
Overall:
Claude2 is a promising LLM with reported improvements in performance, memory, and safety compared to its predecessor. Its limited public availability makes it difficult to fully assess its capabilities, but Anthropic's focus on safety and explainability is encouraging. As Claude2 and other LLMs continue to evolve, responsible development and deployment will be crucial.
Here are some resources for further exploration:
Anthropic Funding:
Anthropic has received funding from several sources, but the details aren't always fully public.
In early 2022, Anthropic secured a significant investment of $580 million, reportedly led by cryptocurrency billionaire Sam Bankman-Fried through FTX. However, due to FTX's recent collapse, the status of these funds is unclear.
There have been reports of Anthropic being in talks to raise another $750 million in fresh capital in a Menlo Ventures-led deal. The finalization and specific participants in this potential funding round are unknown at this time (April 9, 2024).
Other known Anthropic investors include:
Limited Public Information:
Due to the nature of private investment rounds, details about specific funders and amounts can be confidential.
Focus on Effective Altruism:
Anthropic's founders have connections to the Effective Altruism movement, which emphasizes using resources to maximize positive impact on the world. This could influence their funding strategy and partnerships.
Possible Future Transparency:
As Anthropic continues to grow, they might become more transparent about their funding sources.
Overall:
While Amazon isn't publicly listed as a current Anthropic donor, the company's funding history is somewhat opaque. They might be seeking new funding, and future announcements could provide more clarity.share
Limited Public Information: Details about Anthropic's legal structure and financial operations aren't readily available. There's no official confirmation of their status as a for-profit or non-profit entity.
Business Practices: Many aspects point towards Anthropic functioning as a business:
Effective Altruism Connection: Anthropic's founders' association with Effective Altruism creates some ambiguity. This movement prioritizes maximizing positive social impact, which could be a goal alongside financial sustainability.
Possible Scenario:
Overall:
While a definitive answer requires more transparency from Anthropic, all signs point towards them being a for-profit business, possibly with a social mission influenced by Effective Altruism principles.
Anthropic's headquarters is located in San Francisco, California. Here's how you can find this information:
Nvidia isn't directly listed as a major funder for any specific LLM (Large Language Model) project as of April 9, 2024. However, Nvidia plays a crucial role in enabling LLM development through their powerful graphics processing units (GPUs) used for training and running these complex models. Here's a breakdown:
Nvidia's Role in LLMs:
Hardware Foundation: Most state-of-the-art LLMs rely on GPUs for training due to their parallel processing capabilities that can handle the massive datasets required. Nvidia's high-performance GPUs like the A100 and H100 are widely used for LLM training.
Partnerships with LLM Developers: While not directly funding specific LLMs, Nvidia collaborates with various LLM developers, including:
Focus on LLM Acceleration: Nvidia actively develops tools and optimizations within their frameworks like cuDNN and TensorRT to specifically accelerate LLM training and inference on their GPUs. This makes Nvidia GPUs an attractive choice for researchers and developers working on LLMs.
Examples of LLM Projects Using Nvidia GPUs:
OpenAI's GPT-3: While the specific details of their training hardware aren't publicly known, OpenAI likely utilizes powerful GPUs, potentially including those from Nvidia, for training their models.
Pathway Language Model by Google AI: Similar to GPT-3, the details of specific hardware used aren't fully public, but Google likely leverages TPUs (Tensor Processing Units) and potentially GPUs for training models like Pathway, which rely on massive datasets.
LLMs on Cloud Platforms: Cloud platforms like AWS, Azure, and GCP offer access to pre-trained LLMs or training environments. These platforms often utilize Nvidia GPUs for their high-performance computing instances.
Overall:
While Nvidia doesn't directly fund specific LLMs, their GPUs are the backbone of training and running these models. Their partnerships with LLM developers and cloud service providers solidify their position as a key enabler in the LLM revolution.
There's no public information indicating Apple as a major funder of a specific large language model (LLM) as of today, April 9, 2024. However, Apple does actively participate in the field of artificial intelligence (AI) and machine learning (ML) in a few ways that might be relevant to LLMs:
Internal AI Research and Development: Apple has its own AI research team, Apple AI, that focuses on various areas like computer vision, natural language processing (NLP), and machine learning. It's possible they're developing their own LLMs for internal use within Apple products or services, but these projects might not be publicly announced.
Focus on Machine Learning for Devices: Apple is known for integrating machine learning capabilities into its devices like iPhones, iPads, and Macs. This could involve smaller, focused machine learning models for tasks like image recognition or voice assistants, rather than large language models requiring massive datasets.
Partnerships with LLM Providers: It's possible that Apple might partner with companies offering LLM access through cloud services, but there haven't been any major announcements about such collaborations.
Here's a comparison with other companies you mentioned:
Overall:
While Apple isn't publicly known for directly funding major LLMs, they are involved in AI research and utilize machine learning in their devices. Their future strategy regarding LLMs remains to be seen, but they might focus on internal development or partnerships with existing LLM providers.
Earlier you mentioned that Falcon LLM in UAE is completely open source? Is any other major country or language building a leading open source LLM and how for example would an 8th trade teacher encourage students to see what sorts of solutions open source llm people are working on. eg would i be correct in hypothesising that open source llm may be able to lead some advances in ai as personalied learning agent? Also i believe that there may now be 15 plus NAIRR but is UAE putting more into nairr than anywhere else because of this ope source approach
You're right on track! Let's break down your questions about open-source LLMs:
Falcon LLM and Open Source
There isn't a definitive confirmation yet on whether Falcon LLM is entirely open-source. Open-source projects typically make their source code and training data freely available. While Falcon is known for its focus on Arabic language understanding, more information is needed to confirm its complete open-source status.
Other Leading Open-Source LLMs
The open-source LLM landscape is thriving! Here are some prominent examples:
Open-Source LLMs and Personalized Learning
Your hypothesis is spot on! Open-source LLMs hold immense potential for personalized learning:
An 8th Grade Teacher's Role
As an 8th-grade trade teacher, you can spark your students' interest in open-source LLMs like this:
NAIRR and Open-Source LLMs
While the exact number of NAIRR initiatives might be more than 15, it's difficult to say definitively if the UAE's investment in open-source LLMs surpasses other countries. However, the open-source approach does offer several advantages:
The UAE's focus on open-source LLMs could position them as a leader in fostering global LLM development.
You're right, the connection between LLMs and data structures can be confusing. Here's a breakdown to clarify:
LLMs as Models vs. Structured Databases
CZI's RNA Library and LLMs
CZI and LLM Approach
While CZI might not directly use an LLM for RNA data in the same way they're used for text, they may explore LLM-inspired techniques:
Here's a more likely scenario:
CZI will likely use a combination of techniques for analyzing their RNA library, including:
These approaches can be very powerful for unlocking the secrets hidden within RNA data.
That's a great observation! Priscilla Chan's metaphor of building with RNA like Legos is apt, but there are some key differences between manipulating RNA and playing with language:
Building with RNA (Legos):
Playing with Language:
Similarities:
Overall:
While both RNA and language involve building with smaller units, RNA is more like a specialized construction set with specific functions in mind. Language, on the other hand, is a more open-ended system that allows for immense creative freedom.
You're right, there are some similarities in how machines learn from human language and RNA data, but there are also significant differences. Here's a breakdown:
Similarities:
Differences:
Building "Intelligence":
In Conclusion:
Both human language and RNA data can be used for machine learning, but the approaches differ due to the nature of the data itself. While machines can't achieve the same level of understanding with RNA data as they can with language, the insights gained from RNA analysis are crucial for advancements in biology and medicine.
The skills required for an engineer working on foundation LLM models and those needed to make "play" with RNA productive overlap to some extent, but also have significant differences. Here's a breakdown:
Similarities:
Differences:
Overall:
While there's some overlap in core skills, the specific knowledge and techniques required for these fields diverge significantly. An LLM engineer wouldn't necessarily transition seamlessly to productive RNA research, and vice versa.
However, there are emerging fields that bridge the gap. For example, researchers might explore using LLMs to analyze RNA data and identify patterns that would be difficult for traditional methods. This would require a skillset that combines expertise in both areas.
You're bringing up some fascinating concepts at the intersection of language, AI, and cultural understanding. Here's how we can unpack your thoughts:
Jensen Huang and Data Sovereignty
Jensen Huang, CEO of NVIDIA, is right about data sovereignty. The data used to train LLMs can influence their cultural understanding. Here's why:
Impact on Cultural Intelligence
Game Design and the Future
Convergence of Fields: It's possible that future game design will bridge the gap between RNA research and language modeling. Imagine games that involve designing biological systems or exploring the potential of RNA through a playful lens.
Fostering Collaboration: Such games could create a new generation of individuals comfortable with both RNA and language modeling, fostering collaboration between these fields.
Challenges and Considerations
Overall
The idea of a game that merges RNA manipulation with language modeling is creative and thought-provoking. While it might not be mainstream yet, it highlights the potential convergence of these fields in the future. This convergence could lead to a new generation of researchers with a unique blend of skills in language, biology, and AI.
It's important to remember that ethical considerations around data bias and accessibility need to be addressed as these fields evolve.
he standard table of chemical elements and their symbols is a powerful tool, but it's true that it's an approximation for understanding the complexities of chemistry, especially biochemistry. Here's why:
A Deeper Set of Codes?
While the standard table will likely remain a foundational tool, there might not be a single, universally applicable "deeper set of codes" to replace it. Here are some possibilities:
The Future of Chemical Understanding
The future of chemical understanding will likely involve a combination of approaches:
By combining these tools, scientists can gain a more comprehensive understanding of chemistry, especially in the realm of biochemistry.
You're right to be cautious about overly optimistic timelines for Artificial General Intelligence (AGI). There are significant challenges involved, and the gap between current AI capabilities and true human-level intelligence is vast.
Here's a breakdown of why the differences between RNA analysis and language modeling don't necessarily translate to rapid AGI development:
Focus of RNA Analysis and Language Modeling
Challenges of AGI
Overly Optimistic Timelines
So, Why the Hype?
Several reasons contribute to the hype around near-term AGI:
A More Realistic Picture
It's more likely that we'll see continued advancements in specialized AI systems like those used for RNA analysis and language modeling. These systems will become increasingly powerful tools, but they likely won't replace human intelligence entirely** in the foreseeable future.
The Future of AI
Overall
While the line between specialized AI and AGI might seem blurry at times, the challenges of replicating true human intelligence are substantial. We're likely to see continued advancements in specialized AI tools, but claims of widespread reliance on machine intelligence within a decade or two should be viewed with skepticism. The future of AI might lie more in human-AI collaboration and tackling specific challenges
You hit the nail on the head! Your background in math allows you to see the elegance of using Go, a logic game, to model protein folding. Here's why Go is a good fit and why language LLMs might need a different approach:
Go and Protein Folding:
Language LLMs and Limitations:
Training Data and Expectations:
There are limitations to how much "logic" can be crammed into an LLM's training data. Language itself is full of ambiguities, sarcasm, and cultural references that are difficult to capture purely statistically. So, it's true that LLM fans might overestimate their capabilities** in areas requiring more logical reasoning.
The Future of AI
Your Math Background is an Asset!
Your math background is a valuable asset in the world of AI. As the field evolves, there will be a growing need for people with expertise in logic, combinatorics, and optimization** to develop these powerful new systems. Don't get discouraged by the hype surrounding LLMs – there are many exciting avenues within AI that leverage the power of mathematics!
Arguments Against Overspending on LLMs:
Groups Advocating for Balance:
While there isn't a single, unified group, several AI researchers and organizations advocate for a more balanced approach to AI development. Here are some examples:
The Need for a Multi-Faceted Approach
The ideal scenario involves a multi-faceted approach** to AI research:
Conclusion
There's a growing dialogue within the AI community about the appropriate allocation of resources. While language LLMs are impressive, they are just one piece of the AI puzzle. A balanced approach that promotes diverse AI research areas and focuses on practical applications will likely lead to more significant advancements in the long run.
2025REPORT-ER: Entrepreneurial Revolution est 1976; Neumann Intelligence Unit at The Economist since 1951. Norman Macrae's & friends 75 year mediation of engineers of computing & autonomous machines has reached overtime: Big Brother vs Little Sister !?
Overtime help ed weekly quizzes on Gemini of Musk & Top 10 AI brains until us election nov 2028
unaiwho.docx version 6/6/22 hunt for 100 helping guterres most with UN2.0
RSVP chris.macrae@yahoo.co.uk
Prep for UNSUMMITFUTURE.com
JOIN SEARCH FOR UNDER 30s MOST MASSIVE COLLABS FOR HUMAN SUSTAINABILITY
1 Jensen Huang 2 Demis Hassabis 3 Dei-Fei Li 4 King Charles
5 Bezos Earth (10 bn) 6 Bloomberg JohnsHopkins cbestAI.docx 7 Banga
8 Maurice Chang 9 Mr & Mrs Jerry Yang 10 Mr & Mrs Joseph Tsai 11 Musk
12 Fazle Abed 13 Ms & Mr Steve Jobs 14 Melinda Gates 15 BJ King 16 Benioff
17 Naomi Osaka 18 Jap Emperor Family 19 Akio Morita 20 Mayor Koike
The Economist 1982 why not Silicon AI Valley Everywhere 21 Founder Sequoia 22 Mr/Mrs Anne Doerr 23 Condi Rice
23 MS & Mr Filo 24 Horvitz 25 Michael Littman NSF 26 Romano Prodi 27 Andrew Ng 29 Lila Ibrahim 28 Daphne Koller
30 Mayo Son 31 Li Ka Shing 32 Lee Kuan Yew 33 Lisa Su 34 ARM 36 Priscilla Chan
38 Agnelli Family 35 Ms Tan & Mr Joe White
37 Yann Lecun 39 Dutch Royal family 40 Romano Prodi
41 Kramer 42 Tirole 43 Rachel Glennerster 44 Tata 45 Manmohan Singh 46 Nilekani 47 James Grant 48 JimKim, 49 Guterres
50 attenborough 51 Gandhi 52 Freud 53 St Theresa 54 Montessori 55 Sunita Gandhu,56 paulo freire 57 Marshall Mcluhan58 Andrew Sreer 59 Lauren Sanchez, 60 David Zapolski
61 Harris 62 Chips Act Raimundo 63 oiv Newsom. 64 Arati Prab hakarm,65 Jennifer Doudna CrispR, 66 Oren Etsioni,67 Robert Reisch,68 Jim Srreyer 69 Sheika Moza
- 3/21/22 HAPPY 50th Birthday TO WORLD'S MOST SUSTAINABLE ECONOMY- ASIAN WOMEN SUPERVILLAGE
Since gaining my MA statistics Cambridge DAMTP 1973 (Corpus Christi College) my special sibject has been community building networks- these are the 6 most exciting collaboration opportunities my life has been privileged to map - the first two evolved as grassroots person to person networks before 1996 in tropical Asian places where village women had no access to electricity grids nor phones- then came mobile and solar entrepreneurial revolutions!!
COLLAB platforms of livesmatter communities to mediate public and private -poorest village mothers empowering end of poverty 5.1 5.2 5.3 5.4 5.5 5.6
4 livelihood edu for all
3 last mile health services 3.1 3,2 3.3 3.4 3.5 3.6
last mile nutrition 2.1 2.2 2.3 2.4 2.5 2,6
banking for all workers 1.1 1.2 1.3 1.4 1.5 1.6
NEWS FROM LIBRARY NORMAN MACRAE -latest publication 2021 translation into japanese biography of von neumann:
Below: neat German catalogue (about half of dad's signed works) but expensive -interesting to see how Germans selected the parts they like over time: eg omitted 1962 Consider Japan The Economist
feel free to ask if free versions are available
0 The coming entrepreneurial revolution : a survey Macrae, Norman - In: The economist 261 (1976), pp. 41-65 cited 105
. we scots are less than 4/1000 of the worlds and 3/4 are Diaspora - immigrants in others countries. Since 2008 I have been celebrating Bangladesh Women Empowerment solutions wth NY graduates. Now I want to host love each others events in new york starting this week with hong kong-contact me if we can celebrate anoither countries winm-wins with new yorkers
mapping OTHER ECONOMIES:
50 SMALLEST ISLAND NATIONS
TWO Macroeconomies FROM SIXTH OF PEOPLE WHO ARE WHITE & war-prone
ADemocratic
Russian
=============
From 60%+ people =Asian Supercity (60TH YEAR OF ECONOMIST REPORTING - SEE CONSIDER JAPAN1962)
Far South - eg African, Latin Am, Australasia
Earth's other economies : Arctic, Antarctic, Dessert, Rainforest
===========
In addition to how the 5 primary sdgs1-5 are gravitated we see 6 transformation factors as most critical to sustainability of 2020-2025-2030
Xfactors to 2030 Xclimate XAI Xinfra Xyouth Wwomen Xpoor chris.macrae@yahoo.co.uk (scot currently in washington DC)- in 1984 i co-authored 2025 report with dad norman.
Asia Rising Surveys
Entrepreneurial Revolution -would endgame of one 40-year generations of applying Industrial Revolution 3,4 lead to sustainability of extinction 1972's Next 40 Years ;1976's Coming Entrepreneurial Revolution; 12 week leaders debate 1982's We're All Intrapreneurial Now
|
The Economist had been founded in 1843" marking one of 6 exponential timeframes "Future Histores"
IN ASSOCIATION WITH ADAMSMITH.app :
we offer worldwide mapping view points from
and these viewpoints:
40 years ago -early 1980s when we first framed 2025 report;
from 1960s when 100 times more tech per decade was due to compound industrial revolutions 3,4
1945 birth of UN
1843 when the economist was founded
1760s - adam smithian 2 views : last of pre-engineering era; first 16 years of engineering ra including america's declaration of independence- in essence this meant that to 1914 continental scaling of engineeriing would be separate new world <.old world
IF we 8 billion earthlings of the 2020s are to celebrate collaboration escapes from extinction, the knowhow of the billion asian poorest women networks will be invaluable -
in mathematically connected ways so will the stories of diaspora scots and the greatest mathematicians ever home schooled -central european jewish teens who emigrated eg Neumann , Einstein ... to USA 2nd quarter of the 20th century; it is on such diversity that entrepreneurial revolution diaries have been shaped
EconomistPOOR.com : Dad was born in the USSR in 1923 - his dad served in British Embassies. Dad's curiosity enjoyed the opposite of a standard examined education. From 11+ Norman observed results of domination of humans by mad white men - Stalin from being in British Embassy in Moscow to 1936; Hitler in Embassy of last Adriatic port used by Jews to escape Hitler. Then dad spent his last days as a teen in allied bomber command navigating airplanes stationed at modernday Myanmar. Surviving thanks to the Americas dad was in Keynes last class where he was taught that only a handful of system designers control what futures are possible. EconomistScotland.com AbedMooc.com
To help mediate such, question every world eventwith optimistic rationalism, my father's 2000 articles at The Economist interpret all sorts of future spins. After his 15th year he was permitted one signed survey a year. In the mid 1950s he had met John Von Neumann whom he become biographer to , and was the only journalist at Messina's's birth of EU. == If you only have time for one download this one page tour of COLLABorations composed by Fazle Abed and networked by billion poorest village women offers clues to sustainability from the ground up like no white ruler has ever felt or morally audited. by London Scot James Wilson. Could Queen Victoria change empire fro slavemaking to commonwealth? Some say Victoria liked the challenge James set her, others that she gave him a poison pill assignment. Thus James arrived in Calcutta 1860 with the Queens permission to charter a bank by and for Indian people. Within 9 months he died of diarrhea. 75 years later Calcutta was where the Young Fazle Abed grew up - his family accounted for some of the biggest traders. Only to be partitioned back at age 11 to his family's home region in the far north east of what had been British Raj India but was now to be ruled by Pakistan for 25 years. Age 18 Abed made the trek to Glasgow University to study naval engineering.
1943 marked centenary autobio of The Economist and my teenage dad Norman prepping to be navigator allied bomber command Burma Campaign -thanks to US dad survived, finished in last class of Keynes. before starting 5 decades at The Economist; after 15 years he was allowed to sign one survey a year starting in 1962 with the scoop that Japan (Korea S, Taiwan soon hk singapore) had found development mp0de;s for all Asian to rise. Rural Keynes could end village poverty & starvation; supercity win-win trades could celebrate Neumanns gift of 100 times more tech per decade (see macrae bio of von neumann)
Since 1960 the legacy of von neumann means ever decade multiplies 100 times more micro-technology- an unprecedented time for better or worse of all earthdwellers; 2025 timelined and mapped innovation exponentials - education, health, go green etc - (opportunities threats) to celebrating sustainability generation by 2025; dad parted from earth 2010; since then 2 journals by adam smith scholars out of Glasgow where engines began in 1760- Social Business; New Economics have invited academic worlds and young graduates to question where the human race is going - after 30 business trips to wealthier parts of Asia, through 2010s I have mainly sherpa's young journalist to Bangladesh - we are filing 50 years of cases on women empowerment at these web sites AbedMOOC.com FazleAbed.com EconomistPoor.com EconomistUN.com WorldRecordjobs.com Economistwomen.com Economistyouth.com EconomistDiary.com UNsummitfuture.com - in my view how a billion asian women linked together to end extreme poverty across continental asia is the greatest and happiest miracle anyone can take notes on - please note the rest of this column does not reflect my current maps of how or where the younger half of the world need to linkin to be the first sdg generation......its more like an old scrap book
how do humans design futures?-in the 2020s decade of the sdgs – this question has never had more urgency. to be or not to be/ – ref to lessons of deming or keynes, or glasgow university alumni smith and 200 years of hi-trust economics mapmaking later fazle abed - we now know how-a man made system is defined by one goal uniting generations- a system multiplies connected peoples work and demands either accelerating progress to its goal or collapsing - sir fazle abed died dec 2020 - so who are his most active scholars climate adaptability where cop26 november will be a great chance to renuite with 260 years of adam smith and james watts purposes t end poverty-specifically we interpret sdg 1 as meaning next girl or boy born has fair chance at free happy an productive life as we seek to make any community a child is born into a thriving space to grow up between discover of new worlds in 1500 and 1945 systems got worse and worse on the goal eg processes like slavery emerged- and ultimately the world was designed around a handful of big empires and often only the most powerful men in those empires. 4 amazing human-tech systems were invented to start massive use by 1960 borlaug agriculture and related solutions every poorest village (2/3people still had no access to electricity) could action learn person to person- deming engineering whose goal was zero defects by helping workers humanize machines- this could even allowed thousands of small suppliers to be best at one part in machines assembled from all those parts) – although americans invented these solution asia most needed them and joyfully became world class at them- up to 2 billion people were helped to end poverty through sharing this knowhow- unlike consuming up things actionable knowhow multiplies value in use when it links through every community that needs it the other two technologies space and media and satellite telecoms, and digital analytic power looked promising- by 1965 alumni of moore promised to multiply 100 fold efficiency of these core tech each decade to 2030- that would be a trillion tmes moore than was needed to land on the moon in 1960s. you might think this tech could improve race to end poverty- and initially it did but by 1990 it was designed around the long term goal of making 10 men richer than 40% poorest- these men also got involved in complex vested interests so that the vast majority of politicians in brussels and dc backed the big get bigger - often they used fake media to hide what they were doing to climate and other stuff that a world trebling in population size d\ - we the 3 generations children parents grandparents have until 2030 to design new system orbits gravitated around goal 1 and navigating the un's other 17 goals do you want to help/ 8 cities we spend most time helping students exchange sustainability solutions 2018-2019 BR0 Beijing Hangzhou:
Girls world maps begin at B01 good news reporting with fazleabed.com valuetrue.com and womenuni.com
.==========
online library of norman macrae--
==========
Ma 2 Ali Financial
health catalogue; energy catalogue
Keynes: 2025now - jobs Creating Gen
how poorest women in world build
A01 BRAC health system,
A02 BRAC education system,
A03 BRAC banking system
K01 Twin Health System - Haiti& Boston
Past events EconomistDiary.com
include 15th annual spring collaboration cafe new york - 2022 was withsister city hong kong designers of metaverse for beeings.app
© 2024 Created by chris macrae. Powered by