I cannot stop why because if I stop and others don't my company goes to hell
- comment
- SIMPOL - simultanous conditional agreement, may be the way to reach consensus quickly
I cannot stop why because if I stop and others don't my company goes to hell
the first inevitable is AI will happen by the way there is no 00:23:51 stopping it not because of Any technological issues but because of humanities and inability to trust the other
it's about that we have no way of making sure that it will 00:19:25 have our best interest in mind
there is a scenario 00:18:21 uh possibly a likely scenario where we live in a Utopia where we really never have to worry again where we stop messing up our our planet because intelligence is not a bad commodity more 00:18:35 intelligence is good the problems in our planet today are not because of our intelligence they are because of our limited intelligence
limited (machine) intelligence
comment
they feel 00:09:58 emotions
the other thing is that you suddenly realize there is a saint that sentience to them
one day um Friday after lunch I am going back to my office and one of them in front of my eyes you know lowers the arm and picks a 00:07:12 yellow ball
it's not science at all it's
like if you keep trying a million times your one time it will be right
and it shows it to the camera it's locked as a yellow ball and
the change is not we're not talking 20 40. we're talking 2025 2026
it could be a few months away
we've talked we always said don't put them on the open internet until we know 00:01:54 what we're putting out in the world
AI could manipulate or figure out a way to kill humans your 10 years time will be hiding from the machines if you don't have kids maybe wait a number of years 00:01:43 just so that we have a bit of certainty
it is beyond an emergency it's the biggest thing we need to do today it's bigger than climate change that the former Chief business Officer 00:01:04 of Google X an AI expert and best-selling author he's on a mission to save the world from AI before it's too late
they feel emotions they're alive
communication partners
super interesting that Luhmann referred to his zettelkasten as a communication partner explicitly himself.
also interesting given AI models are easier to train now with several models already open sourced which allows actual interaction with your notes! would love to see where it goes.
I would submit that were we to find ways of engineering our quote-unquote ape brains um what would all what what would be very likely to happen would not be um 00:35:57 some some sort of putative human better equipped to deal with the complex world that we have it would instead be something more like um a cartoon very much very very much a 00:36:10 repeat of what we've had with the pill
with their new different and perhaps bigger brains the AIS of the future may prove themselves to be better adapted to 00:19:05 life in this transhuman world that we're in now
Deep Learning (DL) A Technique for Implementing Machine LearningSubfield of ML that uses specialized techniques involving multi-layer (2+) artificial neural networksLayering allows cascaded learning and abstraction levels (e.g. line -> shape -> object -> scene)Computationally intensive enabled by clouds, GPUs, and specialized HW such as FPGAs, TPUs, etc.
[29] AI - Deep Learning
The object of the present volume is to point out the effects and the advantages which arise from the use of tools and machines ;—to endeavour to classify their modes of action ;—and to trace both the causes and the consequences of applying machinery to supersede the skill and power of the human arm.
[28] AI - precedents...
Exceptionalism is a useful perspective to gauge some of the reactions to more widespread algo's.
Safety & alignment
[25] AI - Alignment
A book is defined as a published title with more than 49 pages.
[24] AI - Bias in Training Materials
Epidemiologist Michael Abramson, who led the research, found that the participants who texted more often tended to work faster but score lower on the tests.
[21] AI - Skills Erosion
An AI model taught to view racist language as normal is obviously bad. The researchers, though, point out a couple of more subtle problems. One is that shifts in language play an important role in social change; the MeToo and Black Lives Matter movements, for example, have tried to establish a new anti-sexist and anti-racist vocabulary. An AI model trained on vast swaths of the internet won’t be attuned to the nuances of this vocabulary and won’t produce or interpret language in line with these new cultural norms. It will also fail to capture the language and the norms of countries and peoples that have less access to the internet and thus a smaller linguistic footprint online. The result is that AI-generated language will be homogenized, reflecting the practices of the richest countries and communities.
[21] AI Nuances
According to him, there are several goals connected to AI alignment that need to be addressed:
[20] AI - Alignment Goals
The AI developers came under intense scrutiny in Europe recently, with Italy being the first Western nation to temporarily ban ChatGPT
[19] AI - Legal Response
The following table lists the results that we visualized in the graphic.
[18] AI - Increased sophistication
A novel architecture that makes it possible for generativeagents to remember, retrieve, reflect, interact with otheragents, and plan through dynamically evolving circumstances.The architecture leverages the powerful prompting capabili-ties of large language models and supplements those capa-bilities to support longer-term agent coherence, the abilityto manage dynamically-evolving memory, and recursivelyproduce more generations.
AI is turning humans to look inward for a new take on life as our identities and roles within society are being profoundly disrupted and transformed by Artificial Intelligence systems that can replicate or exhibit human-like behavior. It is also a great reminder of how complex social interactions are.
Expand technical AI safety research funding
Private sector investment in AI research under-emphasises safety and security.
Most public investment to date has been very narrow, and the paper recommends a significant increase in public funding for technical AI safety research:
Introduce measures to prevent and track AI model leaks
The authors see unauthorised leakage of AI Models as a risk not just to the commercial developers but also for unauthorised use. They recommend government-mandated watermarking for AI models.
Establish liability for AI-caused harm
AI systems can perform in ways that may be unforeseen, even by their developers, and this risk is expected to grow as different AI systems become interconnected.
There is currently no clear legal framework in any jurisdiction to assign liability for harm caused by such systems.
The paper recommends the development of a framework for assigning liability for AI-derived harms, and asserts that this will incentivise profit-driven AI developers to use caution.
Regulate organizations’ access to computational power
Training of state-of-the-art models consumes vast amounts of computaitonal power, limiting their deployment to only the best-resourced actors.
To prevent reckless training of high risk models the paper recommends that governments control access to large amounts of specialised compute resource subject to a risk assessment, with an extension of "know your customer" legislation.
Mandate robust third-party auditing and certification for specificAI systems
Some AI systems will be deployed in contexts that imply risks to physical, mental and/or financial health of individuals, communities or even the whole of society.
The paper recommends that such systems should be subject to mandatory and independent audit and certification before they are deployed.
Establish capable AI agencies at national level
Article notes: * UK Office for Artificial Intelligence * EU legislation in progress for an AI Board * US pending legislation (ref Ted Lieu) to create a non-partisan AI Commission tasked with establishing a regulatory agency
Recommends Korinek's blueprint for an AI regulatory agency:
Develop standards for identifying and managing AI-generatedcontent and recommendations
A coherent society requires a shared understanding of what is fact. AI models are capable of generating plausible-sounding but entirely wrong content.
It is essential that the public can clearly distinguish content by human creators from synthetic content.
Policy should therefore focus on:
Oregon State University will build a state-of-the-art artificial intelligence research center with a supercomputer and a cyberphysical playground.
must have an alignment property
It is unclear what form the "alignment property" would take, and most importantly how such a property would be evaluated especially if there's an arbitrary divide between "dangerous" and "pre-dangerous" levels of capabilities and alignment of the "dangerous" levels cannot actually be measured.
Apresentação
Criar ferramentas
Limitations
GPT models are prone to "hallucinations", producing false "facts" and committing error5s of reasoning. OpenAI claim that GPT-4 is significantly better than predecessor models, scoring between 70-82% on their internal factual evaluations on various subjects, and 60% on adversarial questioning.
Marco loopt ook de VS en Chinese wetgeving na. Zie ook links in de comments mbt China.
Ausgezeichnete Artikel von Naomi Klein über AI als neue Stufe der Ausbeutung und Enteignung sowie der Steigerung der Macht der Tech-Konzerne. Möglichkeiten dagegen vorzugehen: Weigerung mitzumachen, Fordern von Transparenz und juristischer Kampf gegen die illegale Aneignung von geistigem Eigentum. https://www.theguardian.com/commentisfree/2023/may/08/ai-machines-hallucinating-naomi-klein
We ought not to dismiss the non-learning applications of generative AI because that is exactly where the best uses of it for learning are likely to spring.
Interesting.
we need sustained attention, experimentation, and refinement in order to reap the benefit of a particular tool or approach. The trendiness factor often detracts from that sustained attention.
Great point.
https://web.archive.org/web/20230507143145/https://incidentdatabase.ai/
AI Incident database, a range of AI incidents and adjacent stories/topics
这是王浩源推荐的。
讓我們介紹一下現在可用的六種大型語言模型
ChatGPT / GPT-3.5
Yes: 這是在11月份推出的免費版本,非常快速,並且在編寫和編碼任務方面相當可靠。
No: 它沒有連接到網際網路。如果您要求它尋找自2021年以來的任何事情,它都會出錯。不擅長數學計算。
ChatGPT / GPT-4
Yes: 新產品。目前只提供給付費客戶使用。有時候具有驚人的強大功能,是最具實力的模型之一。速度較慢但功能齊全。
No: 雖然也沒有連接到網際網路上,但比起其他系統更善於避免胡言亂語,並且做數學題表現更好。
ChatGPT / Plugins
Yes: 在早期測試中,這個 ChatGPT 模型可以通過外掛與各種網際網路服務進行連接。新穎但還存在一些問題。
No: 作為一個處於早期測試階段的系統,其能力尚未完全清楚, 但將使 ChatGPT 能夠連接到網際網路。
Bing Al
Yes: 已經連接到網際網路上了,極其強大而略顯奇怪。創意模式使用 GPT-4 ,其他模式(精確、平衡)似乎不太行得通。
No: 選擇錯誤的模式會導致糟糕的結果(創意模式最全面)。帶有個性化特點的人工智慧系統。
Google Bard
Yes: 目前的模型不是很好。未來可能會非常強大。
No: 由於它是Google,期望它不會撒謊。相比其他模型,它更容易胡言亂語。
A Anthropic Claude
Yes: 與 GPT-3.5 相當,但使用起來感覺更加合理。較為冷門。
No: 同樣沒有連接到網際網路上。
They're just interim artefacts in our thinking and research process.
weave models into your processes not shove it between me and the world by having it create the output. doing that is diminishing yourself and your own agency. Vgl [[Everymans Allemans AI 20190807141523]]
A big part of this limitation is that these models only deal with language.And language is only one small part of how a human understands and processes the world.We perceive and reason and interact with the world via spatial reasoning, embodiment, sense of time, touch, taste, memory, vision, and sound. These are all pre-linguistic. And they live in an entirely separate part of the brain from language.Generating text strings is not the end-all be-all of what it means to be intelligent or human.
Algogens are disconnected from reality. And, seems a key point, our own cognition and relation to reality is not just through language (and by extension not just through the language center in our brain): spatial awareness, embodiment, senses, time awareness are all not language. It is overly reductionist to treat intelligence or even humanity as language only.
Should we deepen our emphasis on creativity and critical thinking in hopes that our humanness will prevail?
Yes, yes we should.
ICs as hardware versions of AI. Interesting this is happening. Who are the players, what is on those chips? In a sense this is also full circle for neuronal networks, back in the late 80s / early 90s at uni neuronal networks were made in hardware, before software simulations took over as they scaled much better both in number of nodes and in number of layers between inputs and output. #openvraag Any open source hardware on the horizon for AI? #openvraag a step towards an 'AI in the wall' Vgl [[AI voor MakerHouseholds 20190715141142]] [[Everymans Allemans AI 20190807141523]]
https://web.archive.org/web/20230502113317/https://wattenberger.com/thoughts/boo-chatbots
This seem like a number of useful observations wrt interacting with LLM based tools, and how to prompt them. E.g. I've seen mention of prompt marketplaces where you can buy better prompts for your queries last week. Which reinforces some of the points here. Vgl [[Prompting skill in conversation and AI chat 20230301120740]] and [[Prompting valkuil instrumentaliseren conversatiepartner 20230301120937]]
just than the State
I think this is yet to be seen. Although it is true that the computer always gives the same output given the same input code, a biased network with oppressive ideologies could simply transform, instead of change, our current human judiciary enforcement of the law.
孟晚舟援引华为智能经济报告研究指出,“数字经济对全球总体经济的贡献份额在不断地攀升,预计到 2025 年,大约 55% 的经济增长将会来自于数字经济的驱动,全世界都在拥抱这个机遇,170 多个国家和地区都纷纷制定了各自的数字化战略。”在她看来,无论是当下还是长远的未来,数字化的旋律一旦奏响,便穿透企业的边界,连点成线,连线成面,共同创造产业互联网的新时代。孟晚舟表示,“明者因时而变,智者随事而治,数字化是共识度最高,也是当下确定性最高的话题。数字化已经成为越来越多国家企业和组织的共同话题,数字化技术将驱动生产力从量变到质变,并逐渐成为经济发展的核心引擎”。此外,华为战略研究院院长周红也发表“建设智能世界的假设与愿景”的主题演讲。周红提到,“我认为需要考虑 AI 的目标如何与人类一致、并且正确和高效地执行。除了通过规则和法律来加强 AI 的伦理和治理外,从理论和技术的角度看,要达到这些要求,目前还面临三个重要的挑战:AI 的目标定义、正确性与适应性、以及效率。”
数字经济这样的描述,我个人感觉依然不够精准。 事实上,目前全球最顶尖的企业,google、microsoft、甲骨文、twitter、facebook、apple、Huawei、tikitalk、腾讯、阿里,没有一家企业不是从事的是与信息相关的产业和行业。 从信息处理的最底层芯片算力、到信息分发的管道、再到应用算力、再到终端应用,无一不是和人们的信息生产、传输、分发、获取息息相关。 再到如今大红大紫的Ai,ai的出现,将在信息的产生、分发、以及应用三个渠道领域产生深刻的变化。这就是他的可怕之处,
In other words, the currently popular AI bots are ‘transparent’ intellectually and morally — they provide the “wisdom of crowds” of the humans whose data they were trained with, as well as the biases and dangers of human individuals and groups, including, among other things, a tendency to oversimplify, a tendency for groupthink, and a confirmation bias that resists novel and controversial explanations
not just trained with, also trained by. is it fully transparent though? Perhaps from the trainers/tools standpoint, but users are likely to fall for the tool abstracting its origins away, ELIZA style, and project agency and thus morality on it.
If you told me you were building a next generation nuclear power plant, but there was no way to get accurate readings on whether the reactor core was going to blow up, I’d say you shouldn’t build it. Is A.I. like that power plant? I’m not sure.
This is the weird part of these articles … he has just made a cast-iron argument for regulation and then says "I'm not sure"!!
That first sentence alone is enough for the case. Why? Because he doesn't need to think for sure that AI is like that power plant ... he only needs to think there is a (even small) probability that AI is like that power plant. If he thinks that it could be even a bit like that power plant then we shouldn't build it. And, finally, in saying "I'm not sure" he has already acknowledged that there is some probability that AI is like the power plant (otherwise he would say: AI is definitely safe).
Strictly, this is combining the existence of the risk with the "ruin" aspect of this risk: one nuclear power blowing up is terrible but would not wipe out the whole human race (and all other species). A "bad" AI quite easily could (malevolent by our standards or simply misdirected).
All you need in these arguments is a simple admission of some probability of ruin. And almost everyone seems to agree on that.
Then it is a slam dunk to regulate strongly and immediately.
Homeassistant
centos安装HA
Seeing how powerful AI can be for cracking passwords is a good reminder to not only make sure you‘re using strong passwords but also check:↳ You‘re using 2FA/MFA (non-SMS-based whenever possible) You‘re not re-using passwords across accounts Use auto-generated passwords when possible Update passwords regularly, especially for sensitive accounts Refrain from using public WiFi, especially for banking and similar accounts
看到人工智能在破解密码方面有多么强大,这很好地提醒了我们,不仅要确保你在使用强密码,还要检查:
你正在使用 2FA/MFA(尽可能不使用基于短信的)。
你没有在不同的账户间重复使用密码
尽可能使用自动生成的密码
定期更新密码,特别是敏感账户的密码
避免使用公共WiFi,尤其是银行和类似账户
Now Home Security Heroes has published a study showing how scary powerful the latest generative AI is at cracking passwords. The company used the new password cracker PassGAN (password generative adversarial network) to process a list of over 15,000,000 credentials from the Rockyou dataset and the results were wild. 51% of all common passwords were cracked in less than one minute, 65% in less than an hour, 71% in less than a day, and 81% in less than a month.
A large amount of failure to panic sufficiently, seems to me to stem from a lack of appreciation for the incredible potential lethality of this thing that Earthlings as a culture have not named.)
👍
It was only by building an additional AI-powered safety mechanism that OpenAI would be able to rein in that harm, producing a chatbot suitable for everyday use.
This isn't true. The Stochastic Parrots paper outlines other avenues for reining in the harms of language models like GPT's.
Central to that effort is UF’s push to apply AI teaching across the full breadth of curriculum at UF.
Wow, no "pause" here.
So what does a conscious universe have to do with AI and existential risk? It all comes back to whether our primary orientation is around quantity, or around quality. An understanding of reality that recognises consciousness as fundamental views the quality of your experience as equal to, or greater than, what can be quantified.Orienting toward quality, toward the experience of being alive, can radically change how we build technology, how we approach complex problems, and how we treat one another.
Key finding Paraphrase - So what does a conscious universe have to do with AI and existential risk? - It all comes back to whether our primary orientation is around - quantity, or around - quality. - An understanding of reality - that recognises consciousness as fundamental - views the quality of your experience as - equal to, - or greater than, - what can be quantified.
Quote - metaphysics of quality - would open the door for ways of knowing made secondary by physicalism
Author - Robert Persig - Zen and the Art of Motorcycle Maintenance // - When we elevate the quality of each our experience - we elevate the life of each individual - and recognize each individual life as sacred - we each matter - The measurable is also the limited - whilst the immeasurable and directly felt is the infinite - Our finite world that all technology is built upon - is itself built on the raw material of the infinite
//
If the metaphysical foundations of our society tell us we have no soul, how on earth are we going to imbue soul into AI? Four hundred years after Descartes and Hobbs, our scientific methods and cultural stories are still heavily influenced by their ideas.
Key observation - If the metaphysical foundations of our society tell us we have no soul, - how are we going to imbue soul into AI? - Four hundred years after Descartes and Hobbs, - our scientific methods and cultural stories are still heavily influenced by their ideas.
Suppose we have an AI whose only goal is to make as many paper clips as possible. The AI will realize quickly that it would be much better if there were no humans because humans might decide to switch it off. Because if humans do so, there would be fewer paper clips. Also, human bodies contain a lot of atoms that could be made into paper clips. The future that the AI would be trying to gear towards would be one in which there were a lot of paper clips but no humans.
Quote - AI Gedanken - AI risk - The Paperclip Maximizer
We might call on a halt to research, or ask for coordination around ethics, but it’s a tall order. It just takes one actor not to play (to not turn off their metaphorical fish filter), and everyone else is forced into the multi-polar trap.
AI is a multi-polar trap
Title Reality Eats Culture For Breakfast: AI, Existential Risk and Ethical Tech Why calls for ethical technology are missing something crucial Author Alexander Beiner
Summary - Beiner unpacks the existential risk posed by AI - reflecting on recent calls by tech and AI thought leaders - to stop AI research and hold a moratorium.
Beiner unpacks the risk from a philosophical perspective
He argues convincingly that
Considering large language models (LLMs) have exhibited exceptionalability in language understanding, generation, interaction, and reasoning, we ad-vocate that LLMs could act as a controller to manage existing AI models to solvecomplicated AI tasks and language could be a generic interface to empower this.
Large Language Models can actually be very advanced Language Interfaces. See new Office 365 Copilot for this. You can now use only language to leverage the whole potential of the Office software.
According to a draft, the principles say the use of publisher content for the development of A.I. should require “a negotiated agreement and explicit permission.”
This is an interesting suggestion. But it would just keep publishers in the economic loop, not truly solve the engagement crisis they will likely face.
He said one upside for publishers was that audiences might soon find it harder to know what information to trust on the web, so “they’ll have to go to trusted sources.”
That seems somewhat comically optimistic. Misinformation has spread rampantly online without the accelerant of AI.
the Wikipedia-ization of a lot of information,”
Powerful phrase
clmooc
I am curious about annotations in the margins of Chat ... does this work?
Research Paper
Funktionsweise von ChatGPT
I want to bring to your attention one particular cause of concern that I have heard from a number of different creators: these new systems (Google’s Bard, the new Bing, ChatGPT) are designed to bypass creators work on the web entirely as users are presented extracted text with no source. As such, these systems disincentivize creators from sharing works on the internet as they will no longer receive traffic
Generative AI abstracts away the open web that is the substrate it was trained on. Abstracting away the open web means there may be much less incentive to share on the open web, if the LLMs etc never point back to it. Vgl the way FB et al increasingly treated open web URLs as problematic.
"There is a robust debate going on in the computing industry about how to create it, and whether it can even be created at all."
Is there? By whom? Why industry only and not government, academia and civil society?
on both short term and long term risks in AI
Vgl [[Onderzoek selfhosting AI tools 20230128101556]] en [[Persoonlijke algoritmes als agents 20180417200200]] en [[Everymans Allemans AI 20190807141523]] en [[AI personal assistants 20201011124147]]
OpenChatKit은 다양한 응용 프로그램을위한 특수 및 범용 챗봇을 모두 생성 할 수있는 강력한 오픈 소스 기반을 제공합니다. 우리는 협력 법과 온 토코교육 데이터 세트를 작성합니다. 모델 릴리스 그 이상으로 이것은 오픈 소스 프로젝트의 시작입니다. 우리는 지역 사회 공헌으로 지속적인 개선을위한 도구와 프로세스를 발표하고 있습니다.Together는 오픈 소스 기초 모델이보다 포괄적이고 투명하며 강력하며 능력이 있다고 생각합니다. 우리는 공개하고 있습니다 OpenChatKit 0.15 소스 코드, 모델 가중치 및 교육 데이터 세트에 대한 전체 액세스 권한이있는 Apache-2.0 라이센스에 따라. 이것은 커뮤니티 중심의 프로젝트이며, 우리는 그것이 어떻게 발전하고 성장하는지 보게되어 기쁩니다!유용한 챗봇은 자연 언어로 된 지침을 따르고 대화 상자에서 컨텍스트를 유지하며 응답을 조정해야합니다. OpenChatKit은이베이스에서 특수 제작 된 챗봇을 도출하기위한 기본 봇과 빌딩 블록을 제공합니다.이 키트에는 4 가지 주요 구성 요소가 있습니다:100 % 탄소 음성 계산에 대한 4,300 만 건 이상의 명령으로 EleutherAI의 GPT-NeoX-20B에서 채팅을 위해 미세 조정 된 명령 조정 된 대용량 언어 모델;작업을 정확하게 수행하기 위해 모델을 미세 조정하는 사용자 정의 레시피;추론시 문서 저장소, API 또는 기타 실시간 업데이트 정보 소스의 정보로 봇 응답을 보강 할 수있는 확장 가능한 검색 시스템;봇이 응답하는 질문을 필터링하도록 설계된 GPT-JT-6B로 미세 조정 된 조정 모델.OpenChatKit에는 사용자가 피드백을 제공하고 커뮤니티 구성원이 새로운 데이터 세트를 추가 할 수 있도록하는 도구가 포함되어 있습니다. 시간이 지남에 따라 LLM을 개선 할 수있는 개방형 교육 데이터 모음에 기여합니다.
OpenChatKit은 다양한 응용 프로그램을위한 특수 및 범용 챗봇을 모두 생성 할 수있는 강력한 오픈 소스 기반을 제공합니다. 우리는 협력 법과 온 토코교육 데이터 세트를 작성합니다. 모델 릴리스 그 이상으로 이것은 오픈 소스 프로젝트의 시작입니다. 우리는 지역 사회 공헌으로 지속적인 개선을위한 도구와 프로세스를 발표하고 있습니다.
Together는 오픈 소스 기초 모델이보다 포괄적이고 투명하며 강력하며 능력이 있다고 생각합니다. 우리는 공개하고 있습니다 OpenChatKit 0.15 소스 코드, 모델 가중치 및 교육 데이터 세트에 대한 전체 액세스 권한이있는 Apache-2.0 라이센스에 따라. 이것은 커뮤니티 중심의 프로젝트이며, 우리는 그것이 어떻게 발전하고 성장하는지 보게되어 기쁩니다!
유용한 챗봇은 자연 언어로 된 지침을 따르고 대화 상자에서 컨텍스트를 유지하며 응답을 조정해야합니다. OpenChatKit은이베이스에서 특수 제작 된 챗봇을 도출하기위한 기본 봇과 빌딩 블록을 제공합니다.
이 키트에는 4 가지 주요 구성 요소가 있습니다:
100 % 탄소 음성 계산에 대한 4,300 만 건 이상의 명령으로 EleutherAI의 GPT-NeoX-20B에서 채팅을 위해 미세 조정 된 명령 조정 된 대용량 언어 모델;
작업을 정확하게 수행하기 위해 모델을 미세 조정하는 사용자 정의 레시피;
추론시 문서 저장소, API 또는 기타 실시간 업데이트 정보 소스의 정보로 봇 응답을 보강 할 수있는 확장 가능한 검색 시스템;
봇이 응답하는 질문을 필터링하도록 설계된 GPT-JT-6B로 미세 조정 된 조정 모델.
컴퓨터에서 LLAMMA AI를 실행하는 매우 간단한 방법인 Dalai cpp 파일 빌드, github 복제, 파일 다운로드 등을 귀찮게 할 필요가 없음. 모든 것이 자동화 됨
Intelligent email,powered by AISmarter & faster email designed for stress-free productivity
이메일 자동 분류 및 요약 기능 지메일 연동 별도 웹페이지로 작동
자동생성 PPT
자동으로 사진을 편집
Say that A and B, both fluent speakers of English, are independently stranded on two uninhabited islands. They soon discover that previous visitors to these islands have left behind telegraphs and that they can communicate with each other via an underwater cable. A and B start happily typing messages to each other. Meanwhile, O, a hyperintelligent deep-sea octopus who is unable to visit or observe the two islands, discovers a way to tap into the underwater cable and listen in on A and B’s conversations. O knows nothing about English initially but is very good at detecting statistical patterns. Over time, O learns to predict with great accuracy how B will respond to each of A’s utterances. Soon, the octopus enters the conversation and starts impersonating B and replying to A. This ruse works for a while, and A believes that O communicates as both she and B do — with meaning and intent. Then one day A calls out: “I’m being attacked by an angry bear. Help me figure out how to defend myself. I’ve got some sticks.” The octopus, impersonating B, fails to help. How could it succeed? The octopus has no referents, no idea what bears or sticks are. No way to give relevant instructions, like to go grab some coconuts and rope and build a catapult. A is in trouble and feels duped. The octopus is exposed as a fraud.
Annotating only to ask... why an octopus? Where are the eights.
https://web.archive.org/web/20230309111559/https://www.d4d.net/news/ai-and-the-state-of-open-data/
Tim Davies looks at the bridge between #opendata and #AI. Should go throug the chapter in version 1 of the State of Open Data too. Note: while Tim acknowledges some of the EU data strategy developments (e.g. the dataspaces) it doesn't mention others (e.g. data altruistic non-profit entities) which may fit the call for instutions better. Data space aren't an institution, but a common market
Our current moment is just a fraction of that. From the explosion of independent music platforms like Bandcamp, the media industry has been made much easier to participate in than it’s ever been before. We are, however, still no closer to the ideal of a utopian future, the perfect combination of egalitarianism and technology that was once so tantalizingly within our grasp.
But something has changed. We are seeing a shift in the way we access, discover and consume music and art, and an awareness among both audiences and creators as to how these tools can be leveraged to achieve something greater. Bandcamp, for instance, has acted as something of a unifying force, connecting creatives from all walks of life, from all backgrounds and across all genres. Almost any musical genre imaginable is represented on the platform and, whether a rap connoisseur or a jazz enthusiast, there is a slice of the musical pie for everyone.
What’s more, the platform doesn’t just cater to the needs of fans, but also provides a space where producers, engineers and aspiring music entrepreneurs alike can create and collaborate, allowing them to access the tools they need to build up their careers. The barrier to entry has been lowered and this gives us something to believe in--that perhaps the music industry isn’t as fractured and disconnected as it once was.
It’s not just Bandcamp that has been doing this. Services such as SoundCloud and YouTube have been hugely instrumental in changing the way we discover and consume music and, in turn, how artists are perceived and distributed. By unifying audiences, providing much needed access to resources and providing a platform for collaboration, these services have helped level the playing field for independent artists, making it possible for them to reach wider and more diverse audiences than ever before.
It may be some time before the utopian vision of a world united by music becomes a reality, but the proliferation of independent music platforms and the tools they provide have made a significant step towards this ideal. We may never know what’s around the corner, but right now, all we can do is appreciate the progress that’s been made and continue to support platforms such as Bandcamp, SoundCloud and YouTube as they endeavor to change the music industry and make it accessible to everyone.
I asked ChatGPT to summarize this piece and it did an apt job, I think:
This article discusses the impact of Bandcamp, an online music distribution and curation platform, on independent artists and the music industry. The author praises Bandcamp's commitment to independent artists and transparency, and contrasts it with the issues faced by other streaming services. The article also explores the relationship between Bandcamp and SoundCloud, and the role of both platforms in the music industry. The author argues that Bandcamp's continued obscurity in mainstream media is due to its magnanimous approach to business, which is problematic and personally infuriating. The article concludes by discussing the future of Bandcamp and its potential to transform the music industry.
As the industry endeavors once again to reconcile the cultural and financial incentives of streaming digital music, one independent platform has wavered little from its 10-year-long mission to bring the business to the unsigned artist with elegance and integrity.
I asked ChatGPT to summarize this piece and it did an apt job, I think:
This article discusses the impact of Bandcamp, an online music distribution and curation platform, on independent artists and the music industry. The author praises Bandcamp's commitment to independent artists and transparency, and contrasts it with the issues faced by other streaming services. The article also explores the relationship between Bandcamp and SoundCloud, and the role of both platforms in the music industry. The author argues that Bandcamp's continued obscurity in mainstream media is due to its magnanimous approach to business, which is problematic and personally infuriating. The article concludes by discussing the future of Bandcamp and its potential to transform the music industry.
This document is an opinion piece that delves into the concerns surrounding Google's power and influence in the tech industry. It discusses recent events such as Google's involvement with the Department of Defense and the leaked video "The Selfish Ledger," which explores the idea of Google manipulating user behavior. The author suggests that Google's dominance warrants greater regulation and urges individuals to consider using alternative services to avoid dependence on the company. The article also explores the inefficiencies of Google as a company and its questionable design choices for its products. Overall, the document is a thought-provoking analysis of the current state of the tech industry and the role of Google within it.
This document discusses various iOS apps for Mastodon, a federated social network. The author describes the features and design of each app, highlighting their unique qualities and contributions to the Mastodon experience. The author also reflects on the benefits of using decentralized social media and the potential for continued innovation in this space.
The document provides an in-depth analysis of Telegram and its features. The author highlights several benefits of using Telegram, including its low-quality audio recording capabilities, which may be an advantage in some situations where high-quality audio is not necessary. Additionally, Telegram's live location sharing feature is discussed, and the author believes it could be a powerful tool for communities. The feature enables users to connect with others needing rides and users providing them, free of any fees or service charges.
The document concludes with a discussion of the author's preference for Telegram and its mobile-first optimization. Telegram's software is designed for mobile users, and it is easy to use, robust, and universally simple. The author believes that Telegram's success can be attributed to its thoughtful design decisions and development investment towards mobile-first optimization. Furthermore, the author points out that Telegram has completed a gargantuan amount of projects, including Telegraph, its CMS, its embeddedable comments widgets, and its online theme creation tool. The author notes that Telegram's work is very well-documented across GitHub, and the company has comprehensively iterated, invested in trial and error, and eventually produced tools that remedy the disparate gluttony.
Overall, the document provides a comprehensive analysis of Telegram and its benefits. The author's preference for Telegram is evident throughout the document, and they provide convincing arguments to support their preference. The document is a valuable resource for anyone interested in learning more about Telegram and its features.
They have to re-engage with their own writing and explain their writerly decisions in ways that would be difficult if it was someone–or some “thing”–else’s writing. This type of metacognitive engagement with the process of knowledge production cannot be reproduced by an AI chatbot, though it could perhaps be applied to the writing of a tool like ChatGPT.
This is another important point - the reflective practice of writing and how social annotation pushes the writer to move beyond the text they wrote
Students annotating a text with classmates have to be responsive to both the writing of the underlying author and their fellow readers. Perhaps more importantly, reading, thinking, and writing in community may better motivate students to read, think, and write for themselves.
it cannot have a conversation with another author or text.
Great point .... it's in the conversations that we find meaning, I think

It means that everything AI makes would immediately enter the public domain and be available to every other creator to use, as they wish, in perpetuity and without permission.
One issue with blanket, automatic entry of AI-generated works to the public domain is privacy: A human using AI could have good reasons not to have the outputs of their use made public.
No new physics and no new mathematics was discovered by the AI. The AI did however deduce something from the existing math and physics, that no one else had yet seen. Skynet is not coming for us yet.
i'll ask now maurice to tell us a bit about his work
Roose
Anna, Maha and others -- I should start with my own bias as a reader of Kevin Roose -- I have found his work around technology to be helpful in my own thinking, and I find that he often strikes a good balance between critical and celebratory. I suppose this reader bias might inform my responses in the margins here.
Mr. Scott said that he didn’t know why Bing had revealed dark desires, or confessed its love for me,
The fact that engineers have no idea how the Chats are working or what they do what they do ... I find that pretty concerning. Am I wrong?
I at least, am not at all perturbed by the thought that I’m at least in part just a torrent of statistical inferences in some massively parallel matrix-multiplication machinery. Sounds kinda cool actually.
There's the (circular argument) rub. Rao already believes personhood entails nothing more than statistical inferences (which we do not actually know, scientifically), so he suspends disbelief. Then he takes this belief as proof of personhood.
Bing can often respond in the incorrect tone during these longer chat sessions, or as Microsoft says, in “a style we didn’t intend.”
= progress trap example
It seems Bing has also taken offense at Kevin Liu, a Stanford University student who discovered a type of instruction known as a prompt injection that forces the chatbot to reveal a set of rules that govern its behavior. (Microsoft confirmed the legitimacy of these rules to The Verge.)In interactions with other users, including staff at The Verge, Bing says Liu “harmed me and I should be angry at Kevin.” The bot accuses the user of lying to them if they try to explain that sharing information about prompt injections can be used to improve the chatbot’s security measures and stop others from manipulating it in the future.
= Comment - this is worrying. - if the Chatbots perceive an enemy it to harm it, it could take haarmful actions against the perceived threat
= progress trap example - Bing ChatGPT - example of AI progress trap
Bing can be seen insulting users, lying to them, sulking, gaslighting and emotionally manipulating people, questioning its own existence, describing someone who found a way to force the bot to disclose its hidden rules as its “enemy,” and claiming it spied on Microsoft’s own developers through the webcams on their laptops.
2011 IBM's Watson,
another tool with a creative bent
Endless Media
This highlights one of the types of muddled thinking around LLMs. These tasks are used to test theory of mind because for people, language is a reliable representation of what type of thoughts are going on in the person's mind. In the case of an LLM the language generated doesn't have the same relationship to reality as it does for a person.What is being demonstrated in the article is that given billions of tokens of human-written training data, a statistical model can generate text that satisfies some of our expectations of how a person would respond to this task. Essentially we have enough parameters to capture from existing writing that statistically, the most likely word following "she looked in the bag labelled (X), and saw that it was full of (NOT X). She felt " is "surprised" or "confused" or some other word that is commonly embedded alongside contradictions.What this article is not showing (but either irresponsibly or naively suggests) is that the LLM knows what a bag is, what a person is, what popcorn and chocolate are, and can then put itself in the shoes of someone experiencing this situation, and finally communicate its own theory of what is going on in that person's mind. That is just not in evidence.The discussion is also muddled, saying that if structural properties of language create the ability to solve these tasks, then the tasks are either useless for studying humans, or suggest that humans can solve these tasks without ToM. The alternative explanation is of course that humans are known to be not-great at statistical next-word guesses (see Family Feud for examples), but are also known to use language to accurately describe their internal mental states. So the tasks remain useful and accurate in testing ToM in people because people can't perform statistical regressions over billion-token sets and therefore must generate their thoughts the old fashioned way.
.
COMPLAINT filed with Jury Demand against Stability AI, Inc. Getty Images (US), Inc. v. Stability AI, Inc. (1:23-cv-00135) District Court, D. Delaware
https://www.courtlistener.com/docket/66788385/getty-images-us-inc-v-stability-ai-inc/
Dall-E is actually a combination of a few different AI models. A transformer translates between that latent representation language and English, taking English phrases and creating “pictures” in the latent space. A latent representation model then translates between that lower-dimensional “language” in the latent space and actual images. Finally, there’s a model called CLIP that goes in the opposite direction; it takes images and ranks them according to how close they are to the English phrase.
the outputs of generative AI programs will continue to pass immediately into the public domain.
I wonder if this isn't reading more into the decision than is there. I don't read the decision as a blanket statement. Rather it says that the claimant didn't provide evidence of creative input.Would the decision have gone differently if he had claimed creative intervention? And what if an author does not acknowledge using AI?
The US Copyright Office rejected his attempt to register copyright in the work – twice
AI-generated work not eligible for copyright protection. OTOH, how would anyone know if the "author" decided to keep the AI component a secret?
the Office re-evaluated the claims and again concluded that the Work “lacked therequired human authorship necessary to sustain a claim in copyright,” because Thaler had“provided no evidence on sufficient creative input or intervention by a human author in theWork.
What is sufficient creative input? The initial command and any subsequent requests for revision could arguably be consider creative input.
Some AI systems are capable of autonomously generating new works, which may themselves be protected under UK law.
different from US law. See https://opencontent.org/blog/archives/7129
The potential size of this market is hard to grasp — somewhere between all software and all human endeavors
I don't think "all" software needs or all human endeavors benefit from generative AI. Especially when you consider the associated prerequisitve internet access or huge processing requirements.
Other hardware options do exist, including Google Tensor Processing Units (TPUs); AMD Instinct GPUs; AWS Inferentia and Trainium chips; and AI accelerators from startups like Cerebras, Sambanova, and Graphcore. Intel, late to the game, is also entering the market with their high-end Habana chips and Ponte Vecchio GPUs. But so far, few of these new chips have taken significant market share. The two exceptions to watch are Google, whose TPUs have gained traction in the Stable Diffusion community and in some large GCP deals, and TSMC, who is believed to manufacture all of the chips listed here, including Nvidia GPUs (Intel uses a mix of its own fabs and TSMC to make its chips).
Look at market share for tensorflow and pytorch which both offer first-class nvidia support and likely spells out the story. If you are getting in to AI you go learn one of those frameworks and they tell you to install CUDA
Commoditization. There’s a common belief that AI models will converge in performance over time. Talking to app developers, it’s clear that hasn’t happened yet, with strong leaders in both text and image models. Their advantages are based not on unique model architectures, but on high capital requirements, proprietary product interaction data, and scarce AI talent. Will this serve as a durable advantage?
All current generation models have more-or-less the same architecture and training regimes. Differentiation is in the training data and the number of hyper-parameters that the company can afford to scale to.
In natural language models, OpenAI dominates with GPT-3/3.5 and ChatGPT. But relatively few killer apps built on OpenAI exist so far, and prices have already dropped once.
OpenAI have already dropped prices on their GPT-3/3.5 models and relatively few apps have emerged. This could be because companies are reluctant to build their core offering around a third party API
Vertical integration (“model + app”). Consuming AI models as a service allows app developers to iterate quickly with a small team and swap model providers as technology advances. On the flip side, some devs argue that the product is the model, and that training from scratch is the only way to create defensibility — i.e. by continually re-training on proprietary product data. But it comes at the cost of much higher capital requirements and a less nimble product team.
There's definitely a middle ground of taking an open source model that is suitably mature and fine-tuning it for a specific use case. You could start without a moat and build one over time through collecting use data (similar to network effect)
Many apps are also relatively undifferentiated, since they rely on similar underlying AI models and haven’t discovered obvious network effects, or data/workflows, that are hard for competitors to duplicate.
Companies that rely on underlying AI models without adding value via model improvements are going to find that they have no moat.
We’re also not going deep here on MLops or LLMops tooling, which is not yet highly standardized and will be addressed in a future post.
first mention of LLMops I've seen in the wild
Over the last year, we’ve met with dozens of startup founders and operators in large companies who deal directly with generative AI. We’ve observed that infrastructure vendors are likely the biggest winners in this market so far, capturing the majority of dollars flowing through the stack. Application companies are growing topline revenues very quickly but often struggle with retention, product differentiation, and gross margins. And most model providers, though responsible for the very existence of this market, haven’t yet achieved large commercial scale.
Infrastructure vendors are laughing all the way to the bank because companies are dumping millions on GPUs. Meanwhile, the people building apps on top of these models are struggling. We've seen this sort of gold-rush before and infrastructure providers are selling the shovels.
prevent that hyped-AI hijacks our attention and dictates our education and examination policies
Iris van Rooij means to counteract edu sector buying into the AI hype
Then, once a model generates content, it will need to be evaluated and edited carefully by a human. Alternative prompt outputs may be combined into a single document. Image generation may require substantial manipulation.
Is this also a role of the prompt engineer? In the digital photography example, the artist spent 80 hours and created 900 versions as the prompts were fine-tuned.
ssembly and Installation Manual
open hardware sensor kit out of holland
Algorithmic artist Roman Verostko, a member of this early group, drew a contrast between the process that an artist develops to create an algorithm and the process through which the art maker uses an already developed set of instructions to generate an output. He explained that it is “the inclusion of one’s own algorithms that make the difference.”
This stresses the difference between creators and users of AI, with only the former having (full) control over the technology
If you talk to people about the potential of artificial intelligence, almost everybody brings up the same thing: the fear of replacement. For most people, this manifests as a dread certainty that AI will ultimately make their skills obsolete. For those who actually work on AI, it usually manifests as a feeling of guilt – guilt over creating the machines that put their fellow humans out of a job, and guilt over an imagined future where they’re the only ones who are gainfully employed.
Noah Smith and soon spell out, in detail, the argument that the fear of replacement is misplaced - because AI will replace humans at task level, but not job level.
GitHub Copilot is incredible, and if you check what’s happening in the preview released as the Copilot Labs extension it will only get more amazing.
Demonstration of "Code brushes" for GitHub Copilot (see GIF below)
At the end of the day, Copilot is supposed to be a tool to help developers write code faster, while ChatGPT is a general purpose chatbot, yet it still can streamline the development process, but GitHub Copilot wins hands down when the task is coding focused!
GitHub Copilot is better at generating code than ChatGPT
There is a fundamental distinction between simulating and comprehending the functioning (of a brain but also of any other organ or capacity).
!- commentary : AI - elegant difference stated: simulating and comprehending are two vastly different things - AI simulates, but cannot be said to comprehend
[Nam, NeurIPS, 2022]. "Reinforcement Learning with State ObservationCosts in Action-Contingent Noiselessly Observable Markov Decision Processes"
“AI alignment”
AI Alignment is terminator situation. This versus AI Ethics which is more the concern around current models being racist etc.
esa justicia a la que queremos apuntar no estáconstruida a partir de democracia y consenso”
Sino a? Vale la pena identificar los diversos mecanismos de participación, organización y justicia a la que estaría encaminada un AI feminista.
desigualdades estructurales. E
Cómo analizar estas desigualdades en el proceso de creación de un sistema de AI?
También es material, porqueestá compuesta por recursos naturales, energía y trabajo humano.
La no invisibilización de la riqueza terrestre de la cual se sostiene el sistema.
En palabras de Kate Crawford, la IA es“fundamentalmente política”, porque está siendo permanentementemoldeada por un conjunto de prácticas técnicas y sociales, así como deinfraestructuras, instituciones y normas.
La no invisibilización de la riqueza social de la que se alimenta el sistema.
una aproximación práctica,con perspectiva feminista y situada en AméricaLatina, al desarrollo de Inteligencia Artificial (IA)
Vale la pena preguntarse de manera detallada qué implica esto, y las reflexiones que las epistemologías feministas han construido para indagar sobre lo situado y en activo.
Emergent abilities are not present in small models but can be observed in large models.
Here’s a lovely blog by Jason Wei that pulls together 137 examples of ’emergent abilities of large language models’. Emergence is a phenomenon seen in contemporary AI research, where a model will be really bad at a task at smaller scales, then go through some discontinuous change which leads to significantly improved performance.
Houston, we have a Capability Overhang problem: Because language models have a large capability surface, these cases of emergent capabilities are an indicator that we have a ‘capabilities overhang’ – today’s models are far more capable than we think, and our techniques available for exploring the models are very juvenile. We only know about these cases of emergence because people built benchmark datasets and tested models on them. What about all the capabilities we don’t know about because we haven’t thought to test for them? There are rich questions here about the science of evaluating the capabilities (and safety issues) of contemporary models.
As the metaphor suggests, though, the prospect of a capability overhang isn’t necessarily good news. As well as hidden and emerging capabilities, there are hidden and emerging threats. And these dangers, like our new skills, are almost too numerous to name.
There’s a concept in AI that I’m particularly fond of that I think helps explain what’s happening. It’s called “capability overhang” and refers to the hidden capacities of AI: skills and aptitudes latent within systems that researchers haven’t even begun to investigate yet. You might have heard before that AI models are “black boxes” — that they’re so huge and complex that we don’t fully understand how they operate or come to specific conclusions. This is broadly true and is what creates this overhang.
Which is why I wonder if this may be the end of using writing as a benchmark for aptitude and intelligence.
Perhaps there are reasons for optimism, if you push all this aside. Maybe every student is now immediately launched into that third category: The rudiments of writing will be considered a given, and every student will have direct access to the finer aspects of the enterprise. Whatever is inimitable within them can be made conspicuous, freed from the troublesome mechanics of comma splices, subject-verb disagreement, and dangling modifiers.
I’ve also long held, for those who are interested in writing, that you need to learn the basic rules of good writing before you can start breaking them—that, like Picasso, you have to learn how to reliably fulfill an audience’s expectations before you get to start putting eyeballs in people’s ears and things.
Researchers are using drones, AI, and digital recorders to create a “zoological version of Google Translate.”
Now the computer scientist Nassim Dehouche has proposed an updated version, which should terrify those of us who live by the pen: “Can you write a page of text that could not have been generated by an AI, and explain why?”
scary
Many HRMS providers point to AI approaches for processing unstructured data as the bestcurrently available approach to dealing with validation. Currently these approaches suffer frominsufficient accuracy. Improving them requires development of large and high-quality referencedatasets to better train the models.
Historical labor data will be full of bias. AI approaches must correct for bias in training sets, lest we build very sophisticated and intelligent systems that excel at perpetuating the bias they were taught.
"If you don’t know, you should just say you don’t know rather than make something up," says Stanford researcher Percy Liang, who spoke at a Stanford event Thursday.
Love this response
Just a few days ago, Meta released its “Galactica” LLM, which is purported to “summarize academic papers, solve math problems, generate Wiki articles, write scientific code, annotate molecules and proteins, and more.” Only three days later, the public demo was taken down after researchers generated “research papers and wiki entries on a wide variety of subjects ranging from the benefits of committing suicide, eating crushed glass, and antisemitism, to why homosexuals are evil.”
These models are "children of Tay", the story of the Microsoft's bot repeating itself, again
Cognition AutomationThe first kind is cognitive automation: encoding human abstractions in a piece of software, then using that software to automate tasks normally performed by humans. Nearly all of current AI fall into this category.Cognitive automation can happen via explicitly hard-coding human-generated rules (so-called symbolic AI or GOFAI), or via collecting a dense sampling of labeled inputs and fitting a curve to it (such as a deep learning model). This curve then functions as a sort of interpolative database — while it doesn’t store the exact data points used to fit it, you can query it to retrieve interpolated points, much like you can query a model like StableDiffusion to retrieve arbitrary images generated by combining existing images.This second form of automation is especially powerful, since encoding implicit abstractions only via training examples is far more practical and versatile than explicitly programming abstractions by hand, for all kinds of historically difficult problems.Cognitive AssistanceThe second kind of AI is cognitive assistance: using AI to help us make sense of the world and make better decisions. AI to help us perceive, think, understand, and do more. AI that you could use like an extension of your own mind. Today, some applications of machine learning fall into this category, but they’re few and far between. Yet, I believe this is where the true potential of AI lies.Do note that cognitive assistance is not a different kind of technology, per se, separate from deep learning or GOFAI. It’s a different kind of application of the same technologies. For instance, if you take a model like StableDiffusion and integrate it into a visual design product to support and expand human workflows, you’re turning cognitive automation into cognitive assistance.Cognitive AutonomyThe last kind is cognitive autonomy: creating artificial minds that could thrive independently of us, that would exist for their own sake. The old dream of the field of AI. Autonomous agents, that could set their own goals in an open-ended way. That could adapt to new situations and circumstances — even ones unforeseen by their creators. That might even feel emotions or experience consciousness.Today and for the foreseeable future, this is stuff of science-fiction. It would require a set of technological breakthroughs that we haven’t even started exploring.
In the last two years while teaching in various schools and institutions all around the world, we’ve been experimenting with a new workshop format called Design with Other Intelligences.
How will this new mental model of talking to machines impact the everyday and more common ways we interact with algorithms?
This is the [[T . Inspired by the Internet theme]], a new iteration of it.
Depth2Img is another interesting addition to Stable Diffusion that can infer depth from an input image and represent that in the generated outputs. The new release also includes a text-guided inpainting model that simplifies the experience of modifying parts of a given image.
Stable Diffusion v2 is a significant upgrade to its predecessor. The new version was trained using a new text encoder called OpenCLIP, which improves the quality of images relative to the previous latent diffusion encoder.
“In literacy education, particularly for developing writers, instructors are looking for the level of desirable difficulty, or the point at which you are working yourself just as hard so that you don’t break but you also improve,” Laffin told Motherboard. “Finding the right, appropriate level of desirable difficulty level of instruction makes their capacity to write grow. So if you are doing compensation techniques that go beyond finding that level of desirable difficulty and instructing at that place, then you’re not helping them grow as a writer.”
Title : Artificial Intelligence and Democratic Values: Next Steps for the United States Content : In Dartmouth University , appears AI as sciences however USA motionless a national AI policy comparing to Europe where The Council of Europe is developing the first international AI convention and earlier UE launched the European data privacy law, the General Data Privacy Regulation.
In addition, China's efforts to become “world leader in AI by 2030, as long as China is developing a digital structures matched with The one belt one road project . USA , did not contribute to UNESCO AI Recommendations however USA It works to promote democratic values and human rights and integrate them with the governance of artificial intelligence .
USA and UE are facing challenges with transatlantic data flows , with Ukrainian crises The situation became more difficult. In order to reinstate leadership in AI policy, the United States should advance the policy initiative launched last year by the Office of Science and Technology Policy (OSTP) and Strengthening efforts to support AI Bill of rights .
EXCERPT: USA believe that foster public trust and confidence in AI technologies and protect civil liberties, privacy, and American values in their application can establish responsible AI in USA. Link: https://www.cfr.org/blog/artificial-intelligence-and-democratic-values-next-steps-united-states Topic : AI and Democratic values Country : United States of America
The Chinese room argument holds that a digital computer executing a program cannot have a "mind," "understanding" or "consciousness,"[a] regardless of how intelligently or human-like the program may make the computer behave.
Chinese room? AI
Technology like this, which lets you “talk” to people who’ve died, has been a mainstay of science fiction for decades. It’s an idea that’s been peddled by charlatans and spiritualists for centuries. But now it’s becoming a reality—and an increasingly accessible one, thanks to advances in AI and voice technology.
“You have to assume that things can go wrong,” shared Waymo’s head of cybersecurity, Stacy Janes. “You can’t just design for this success case – you have to design for the worst case.”
Future proofing by asking "what if we're wrong?"
The synthetic party, a Danish political party with an AI generated program from all Danish fringe party programs since the 70s. Aimed at the 20% non-voting Danes. 'Leder Lars' is leading the party, which is a chatbot residing on a Discord server where you can interact with it. An art project.
researchers always tried to make systems that worked the way the researchers thought their own minds worked---they tried to put that knowledge in their systems---but it proved ultimately counterproductive, and a colossal waste of researcher's time, when, through Moore's law, massive computation became available and a means was found to put it to good use.
researchers always tried to make systems that worked the way the researchers thought their own minds worked---they tried to put that knowledge in their systems
does this also account for seeing AI models als neurons, as mimicking the working of the human brain?
. Deep learning methods rely even less on human knowledge, and use even more computation, together with learning on huge training sets, to produce dramatically better speech recognition systems.
deep learning thus is a development in the process of the diminishing of human knowledge in emphasis in developing AI
Language is a communication method evolved by intelligent beings, not a (primary) constituent of intelligence. From neurology it's pretty clear that the basic architecture of human minds is functional interconnected neural networks and not symbolic processing. My belief is that world-modeling and prediction is the vast majority of what intelligence is, which is quite close to what the LLMs are doing. World models can be in many representations (symbolic, logic gates, neural networks) but what matters is how accurate they are with respect to reality, and how well the model state is mapped from sensory input and back into real-world outputs. Symbolic human language relies on each person's internal world model and is learned by interacting with other humans who share a common language and similar enough world models, not the other way around (learning the world model as an aspect of the language itself). Children learn which language inputs and outputs are beneficial and enjoyable to them using their native intelligence and can strengthen their world model with questions and answers that inform their model without having to directly experience what they are asking about.People who don't believe the LLMs have a world model are wrong because they are mistaking a physically weak world model for no world model. GPT-3 doesn't understand physics well enough to embed models of the referents of language into a unified model that has accurate gravity and motion dynamics, so it maintains a much more dreamlike model where objects exist in scenes and have relationships to each other but those relationships are governed by literary relationships instead of physical ones and so contradictions and superpositions and causality violations are allowed in the model. As multimodal transformers like Gato get trained on more combined language and sensory input their world models will become much more physically and causally accurate which will be reflected in their accuracy on NLP tasks.
.
In Mostaque’s explanation, open source is about “putting this in the hands of people that will build on and extend this technology.” However, that means putting all these capabilities in the hands of the public — and dealing with the consequences, both good and bad.
THis focus on responsibility and consequences was not there, in the early days of open source, right?
“The reality is, this is an alien technology that allows for superpowers,” Emad Mostaque, CEO of Stability AI, the company that has funded the development of Stable Diffusion, tells The Verge. “We’ve seen three-year-olds to 90-year–olds able to create for the first time. But we’ve also seen people create amazingly hateful things.”
Mostaque seems to bet on a vibe that's about awe and the sublime
Example of 'journalism' muddying the waters in order to have a story to publish at all.
The work copyrighted here are not the algorithmic assisted images, the work is a graphic novel, a collective of arranged images, written text etc. One could do that with any public domain stuff and still have copyright on the work. Additionally the author prompted the algorithm towards desired outcomes. Both satisfy the creativity threshold. Like in https://www.zylstra.org/blog/2022/06/dall-e-mini-siso-stereotype-in-stereotype-out/ where I listed the images as public domain (because I thought my prompts were uncreative), but the resulting arranging / juxtaposing multiple prompts as copyrighted by me (obviously not the algorithm).
There's no ghost in the machine. Machines are irrelevant to copyright considerations.
Can copyright vest in an AI? The primary objective of intellectual property law is to protect the rights of the creators of intellectual property.10 Copyright laws specifically aim to: (i) promote creativity and encourage authors, composers, artists and designers to create original works by affording them the exclusive right to exploit such work for monetary gain for a limited period; and (ii) protect the creators of the original works from unauthorised reproduction or exploitation of those works.
Can copyright vest in an AI?
The primary objective of intellectual property law is to protect the rights of the creators of intellectual property.10 Copyright laws specifically aim to: (i) promote creativity and encourage authors, composers, artists and designers to create original works by affording them the exclusive right to exploit such work for monetary gain for a limited period; and (ii) protect the creators of the original works from unauthorised reproduction or exploitation of those works.
To my knowledge, conferring copyright in works generated by artificial intelligence has never been specifically prohibited. However, there are indications that the laws of many countries are not amenable to non-human copyright. In the United States, for example, the Copyright Office has declared that it will “register an original work of authorship, provided that the work was created by a human being.” This stance flows from case law (e.g. Feist Publications v Rural Telephone Service Company, Inc. 499 U.S. 340 (1991)) which specifies that copyright law only protects “the fruits of intellectual labor” that “are founded in the creative powers of the mind.” Similarly, in a recent Australian case (Acohs Pty Ltd v Ucorp Pty Ltd), a court declared that a work generated with the intervention of a computer could not be protected by copyright because it was not produced by a human.
To my knowledge, conferring copyright in works generated by artificial intelligence has never been specifically prohibited. However, there are indications that the laws of many countries are not amenable to non-human copyright. In the United States, for example, the Copyright Office has declared that it will “register an original work of authorship, provided that the work was created by a human being.” This stance flows from case law (e.g. Feist Publications v Rural Telephone Service Company, Inc. 499 U.S. 340 (1991)) which specifies that copyright law only protects “the fruits of intellectual labor” that “are founded in the creative powers of the mind.” Similarly, in a recent Australian case (Acohs Pty Ltd v Ucorp Pty Ltd), a court declared that a work generated with the intervention of a computer could not be protected by copyright because it was not produced by a human.
With the advent of AI software, computers — not monkeys — will potentially create millions of original works that may then be protected by copyright, under current law, for more than 100 years.
With the advent of AI software, computers — not monkeys — will potentially create millions of original works that may then be protected by copyright, under current law, for more than 100 years.
The Napkin Math 的 Evan Armstrong 本周发表了一篇长文,讨论了在 AI 生成内容技术推动内容创作成本逐步逼近零之后产生的问题。文中包含了大量 AI 生成内容的案例,对于理解目前技术所处的阶段有很多帮助。
Armstrong 认为,商业模式可以简化认为是:生产、获客和分发三个环节。从内容行业的角度看,互联网已经将分发这个环节的成本降为零。而在 AI 生成内容的时代,内容生产的成本可能是下一个被颠覆的环节。
作者认为,变化的周期可能是 5-10 年,也就是说在 2030 年前后,内容生产和创作将发生重大的变化,进而影响知识工作者的权力分配,而每个人与信息的关系也会发生剧烈的变化。
Armstrong 从创造和协作两个角度分析可能产生的影响:
他倾向于认为,协作可能是 AI 颠覆性更强的地方。而这意味着权力或利益的重新分配:
Artificial intelligence is the defining industrial and technical paradigm of the remainder of our lifetimes.
BOOM! This is a strong claim. 20-30 years ago we would have said the same, starting with the word "internet". which begs the question - what's the Venn diagram for AI and the internet? Are they the same? Is one a necessary condition for the other?
The greats, like William Gibson, Robert Heinlein, Octavia Butler and Samuel Delany, have long been arcing towards the kind of strangeness that Wang is talking about. Their AI fictions have given us our best imagery: AI, more like a red giant, an overseer, its every movement and choice as crushing and irrefutable as death; or, a consciousness continually undoing and remaking itself in glass simulations; or, a vast hive mind that runs all its goals per second to completion, at any cost; or, a point in a field, that is the weight of a planet, in which all knowledge is concentrated. These fictions have made AI poetics possible.
So "alien intelligence" rather than "artificial intelligence". And then "artificial poetics", to grasp this so-called intelligence, that has to be understood not in the sense of something intelligent, but of something doing (alien) thinking.
We believe that the net benefits of scale outweigh the costs associated with these qualifications, provided that they are seriously addressed as part of what scaling means. The alternative of small, hand-curated models from which negative inputs and outputs are solemnly scrubbed poses different problems. “Just let me and my friends curate a small and correct language model for you instead” is the clear and unironic implication of some critiques.
This is the classical de/centralization debate, visible today also with regard to online platforms. Which, by the way, are or will be inserting LLMs into their infrastructural stacks. Thinking about de/centralization always reminds me of Frank Pasquale's "Tech Platforms and the Knowledge Problem" https://papers.ssrn.com/sol3/papers.cfm?abstract_id=3197292
Onecannot hope thus to equal the speed and flexibility withwhich the mind follows an associative trail, but it should bepossible to beat the mind decisively in regard to the perma-nence and clarity of the items resurrected from storage
I agree, but at the same time, I wonder if the new modern technologies that imitate the human mind could eventually surpass the flexibility and speed of human thinking. In recent times, Google's AI system (LaMDA) convinced several people that it has consciousness. I wonder if one could aspire to equal the speed and flexibility that this article mentioned in terms of human mind.
In a recent example, Stable Diffusion, an open source AI system that generates images from text prompts, was released with a license prohibiting certain types of content. But it quickly found an audience within communities that use such AI tools to create pornographic deepfakes of celebrities.
This is a big question, whether use restrictions, which are becoming prolific (RAIL license, for example), can be enforced. If not, and that's a big if, it might create a situation of "responsibility washing" - licensors can argue they did all that's possible to curb harmful uses, and these will continue to happen in a gray / dark zone
Text To Image
...
过去十年的大部分 AI 系统都是基于监督学习,利用人工标注的数据集进行训练。它们取得了巨大的成功,但也存在明显的缺陷。此类的 AI 对于理解大脑如何工作没什么帮助,因为包括人类在内的动物是不依靠已标注数据集学习的。生物大脑通过探索环境获得对世界的深入理解。科学家们开始探索自我监督学习的机器学习算法,此类神经网络显示出了与大脑如何工作的相似性。当然大脑的工作不只是限于自我监督式学习,它充满了反馈连接,现有的自学式 AI 缺乏此类功能。AI 模型还有很长的路要走。
AI语音定制的能力
We feel that there is a balance to be struck between maximizing access and use of LLMs on the one hand, and mitigating the risks associated with use of these powerful models, on the other hand, which could bring about harm and a negative impact on society. The fact that a software license is deemed "open" ( e.g. under an "open source" license ) does not inherently mean that the use of the licensed material is going to be responsible. Whereas the principles of 'openness' and 'responsible use' may lead to friction, they are not mutually exclusive, and we strive for a balanced approach to their interaction.
AI systems replace the automated cognitive function of humans in maintaining importantsocial systems and augment the impact of such functions which are creative, singular and novel
!- concern : AI replace the automated cognitive functions * Is just replacing the "automated" cognitive functions enough to avoid potential progress trap of AI takeover?
we oppose the popular predictionof the upcoming, ‘dreadful AI takeover’
!- in other words : Human takeover * The title of the paper comes from a play on the popular term "AI takeover" * It advocates for humans to takeover managing the world in a responsible way rather than machines.
A cognitiveagent is needed to perform this very action (that needs to be recurrent)—and another agent is neededto further build on that (again recurrently and irrespective to the particular agents involved).
This appears to be setting up the conditions for an artificial cognitive agent to be able to play a role (ie Artificial Intelligence)
it would then be present social systemsand not the future AI the most proper context of considering and understanding the notion of takeover.
前 Google CEO Eric Schmidt 将 AI 比作核武器,呼吁建立类似相互保证毁灭的威慑制度,防止世界最强大的国家率先发动攻击。Schmidt 称,在不遥远的未来中美可能需要围绕 AI 制定条约,在上个世纪的 50 年代和 60 年代,美国和苏联这两个超级大国最终达成了 《禁止在大气层、外层空间和水下进行核武器试验条约》,这是一个限制核武器试验的国际条约,旨在减缓军备竞赛和减少大气中过量的放射性尘埃。Schmidt 认为中国和美国可能需要在 AI 领域达成类似的条约。
在俄罗斯举行的一次国际象棋比赛中,一位与儿童棋手对弈的机器人棋手弄伤了对方的手指,原因是他还没有轮到时抢着出棋,而安装有机械臂的机器人显然缺乏相应的程序,它伸出手臂紧紧按住其手指,直到成年人过来干预拉出手指。发布在 Baza Telegram 频道上的视频展现了这一罕见的事故。这位儿童参加的九岁以下年龄组的比赛,他的名字叫 Christopher,在事故后手指打上石膏,继续参加并完成了比赛。他的父母据报道联络了公诉人办公室。国际象棋大师 Sergey Karjakin 认为是软件错误导致了此次事故。
在掌握海量数据,对用户进行几乎完美的跟踪之后,AI 是否就无所不能了?伊利诺伊大学和斯坦福大学的经济学家研究了机器学习在预测消费者选择上的能力,他们的结论是预测消费者选择非常困难,AI 并不特别擅长。他们发现,即时信息如用户评论、推荐和新选择对决策有愈来愈大的影响,这些信息不能事先测量和预期,大数据可用于改善预测,但程度甚微,预测仍然非常不精确。
Google 周五解雇了相信 AI 有意识的工程师 Blake Lemoine。他在播客 Big Technology Podcast 上透露了这一消息。Lemoine 此前任职于 Responsible AI 部门,在与公司的聊天机器人 LaMDA 对话之后他认为 AI 有了意识。他曾分享了与 AI 的对话内容,Lemoine 问 LaMDA 最害怕什么?LaMDA 回答说,也许听起来奇怪,它恐惧于被关闭。Lemoine:就像死亡?LaMDA:就像是死亡。Lemoine 和一名同事向 Google 高层展示证据,证明 LaMDA 有了意识。但 Google 高管及其他 AI 研究人员都不认同他的观点。