Tag Archives: AI

Seamless AI text, sound, and video

It’s only 8 seconds long, but this clip of and old sailor could easily be mistaken for real:

And beyond looking real, here is what Google’s new Flow video production platform can do:

Body movement, lip movement, objects moving naturally in gravity, we have the technology to create some truly incredible videos. On the one hand, we have amazing opportunities to be creative and expand the capabilities of our own imaginations. On the other hand we are entering into a world of deep fakes and misinformation.

Such is the case with most technologies. They can be used well and can be used poorly. Those using it well will amaze us with imagery and ideas long stuck in people’s heads without a way previously to express them. Those using it poorly will anger and enrage us. They will confuse us and make it difficult to discern fake news from real.

I am both excited and horrified by the possibilities.

The tail wagging the dog

I recently wrote, ‘The school experience’ where I stated, “I don’t know how traditional schools survive in an era of Artificial Intelligence?” In that post I was focused on removing the kind of things we traditionally do with opportunities to experience learning in the classroom (with and without AI).

What’s interesting about this is that the change will indeed come, but not for the right reasons. The reason we’ll see a transformation of schools happen faster than expected is because with AI being constantly used to do homework, take notes, and do textbook assignments, grades are going to be inflated and it will be hard to discern who gets into universities.

This will encourage two kinds of changes in schools. On the one hand we will see a movement backwards to more traditional testing and reduced innovation. This is the group that wants to promote integrity, but blindly produces students who are good memorizers and are good at wrote learning. However, not producing students ready to live in our innovative and ever-changing world.

The second kind of school will promote competencies around thinking, knowing, and doing things collaboratively and creatively. These are the real schools of the future.

But I wonder which of these schools will universities be more interested in? Which practices will universities use? It’s easier to invigilate an exam that is based on wrote learning than it is to mark group projects in a lecture hall of 200+ students. So what kind of students are universities going to be looking for?

I fear that this might be a case of the tail wagging the dog and that we could see a movement towards ‘traditional learning’ as a pathway to a ‘good’ university… The race to the best marks in a school that tests in traditional ways and has ‘academic rigour’ could be the path that universities push.

This is a mistake.

The worst part of schooling is marks chasing. It undermines meaningful feedback and it misses the point that this is a learning environment with learning opportunities. Instead it’s about the mark. The score that gets averaged into GPA’s and meets minimum requirements to get into programs or schools of choice after high school.

The question I ponder is if universities will continue to focus on that metric and continue to wag the dog in this way, or will they start looking more meaningfully at other metrics like portfolios and presentations? Will they take the time to do the work necessary to really assess the student as a learner, or will they just continue to collect marks chasers and focus on accepting kids who come from schools that are good at differentiating those marks in traditional ways?

This could be an exciting time for universities to lead the way towards truly innovative practices rather than being the last bastion of old ways of teaching and learning… Old ways being perpetuated by a system that values marks over thinking, traditions over progress, and old practices over institutions of truly higher learning.

University entry is the tail wagging the dog, and so the way that universities respond to AI doing work that students have had to do will determine how quickly schools innovate and progress.

The School Experience

I don’t know how traditional schools survive in an era of Artificial Intelligence? There are some key elements of school that are completely undermined by tools that do the work faster and more effectively than students. Here are three examples:

  1. Homework. If you are sending homework such as an essay home, it’s not a question of whether or not a student uses AI, it’s a question of how much AI is being used. Math homework? That’s just practice for AI, not the student.
  2. Note taking. From recording and dictating words to photographing slides and having them automatically transcribed, if a traditional lecture is the format, AI is going to outperform any physical note taking.
  3. Textbook work? Or questions about what happened in a novel? This hunt-and-peck style assignment used to check to see if a student did the reading, but unless it’s a supervised test situation, a kid can get a perfect score without reading a single page.

So what do we want students to do at school? Ultimately it’s about creating experiences. Give them a task that doesn’t involve taking the project home. Give them a task where they need to problem solve in teams. Engage in content with them then have them debate perspectives… even provide them with opportunities to deepen their perspectives with AI before the debate.

Class time is about engaging in and with the content, with each other, and with tools that help students understand and make meaning.. Class isn’t consumption of content, it’s engaging with content, it’s engaging in collaborative challenges, it’s time to be creative problem-solvers.

Don’t mistake the classroom experience with entertaining students, it’s not about replacing the content or the learning with Bill Nye the Science Guy sound bites of content… it’s about creating experiences where students are challenged, while in the class, to solve problems that engage them. And this doesn’t mean avoiding AI, it does mean that it is used or not used with intentionality and purpose.

We need to examine what the school experience looks like in an era when technology makes traditional schooling obsolete. We didn’t keep scribing books after the printing press. Blacksmiths didn’t keep making hand-forged nails after we could mass produce them. Yet AI can efficiently and effectively produce the traditional work we ask for in schools and somehow we want students to mass produce the work the old way?

How do we transform the school experience so that it is meaningful and engaging for students… not AI?

*I used AI (Copilot) to suggest the production of nails as being a redundant item no longer created by blacksmiths. I also use AI to create most of the images on my blog, including the one with this post, with a prompt that took a couple attempts until Copilot offered, “Here comes a fresh take! A Rube Goldberg-style school, where the entire structure itself is a fantastical machine, churning out students like a whimsical knowledge factory.”

Robot dogs on wheels

We seem to have a fascination with robots being more and more like humans. We are training them to imitate the way we walk, pick things up, and even gesture. But I think the thing most people aren’t realizing is how much better than humans robots will be (very soon).

The light bulb went on for me a few months back when a saw a video of a humanoid robot lying on the floor. It bent it’s knees completely backwards, placing it’s feet on either side of it’s hips and lifted itself to standing from close to it’s center of gravity. Then it walked backwards a few steps before rotating it’s body 180º to the direction it was walking. 

I was again reminded of this recently when I saw a robotic dog going over rugged terrain, and when it reached level ground, instead of running it just started to roll on wheels. The wheels were locked into position when the terrain was rougher, and it made more sense to be a dog-like quadruped to maximize mobility. 

There is no reason for a robot to have a knee with the same limited mobility as our knees. A hand might have more functionality with 3 fingers and a thumb, or 4 fingers and 2 opposable thumbs on either side of the fingers. Furthermore, this ‘updated’ hand can have the dexterity to pick something up using either side of it’s hand. It would be like if the hand had two palms, simply articulating finger digits to go the opposite way when it is practical. Beyond fully dexterous hands, we can start to use our imagination: heads that rotate to any direction, a third arm, the ability to run on all-fours, incredible jumping ability, moving faster, being stronger, and viewing everything with 360º cameras that have the ability to magnify an object far beyond human eyesight capabilities. All the while processing more information than we can hold in our brains at once. 

Robot dogs on wheels are just the first step in creating robots that don’t just replicate the mobility and agility of living things, but actually far exceed any currently abilities that we can think of. Limitations to these robots of the near future are only going to be a result of our lack of imagination… human imaginations, because we can’t even know what an AI will think of in 20-30 years. We don’t need to worry about human-like robots, but we really do need to worry about robots that will be capable of things we currently think are impossible… And I think we’ll start to see these in the very near future. The question is, will they help humanity or will they be used in nefarious ways? Are we going to see gun wielding robot dogs or robots performing precision surgery and saving lives? I think both, but hopefully we’ll see more of these amazing robots helping humanity be more human.  

The Right Focus

When I wrote, ‘Google proof vs AI proof‘, I concluded, “We aren’t going to AI proof schoolwork.

While we were successful in Google proofing assignments by creating questions that were not easily answerable using a Google search, we simply can’t make research based questions that will stump a Large Language Model artificial intelligence.

I’ve also previously said that ‘You can’t police it‘. In that post I stated,

“The first instinct with combating new technology is to ban and/or police it: No cell phones in class, leave them in your lockers; You can’t use Wikipedia as a source; Block Snapchat, Instagram, and TikTok on the school wifi. These are all gut reactions to new technologies that frame the problem as policeable… Teachers are not teachers, they aren’t teaching, when they are being police.

It comes down to a simple premise:

Focus on what we ‘should do’ with AI, not what we ‘shouldn’t do’.

Outside the classroom AI is getting used everywhere by almost everyone. From programming and creating scripts to reduce workload, to writing email responses, to planning vacations, to note taking in meetings, to creating recipes. So the question isn’t whether we should use AI in schools but what we should use it for?

The simple line that I see starts with the question I would ask about using an encyclopedia, a calculator, or a phone in the classroom, “How can I use this tool to foster or enhance thinking, rather than have the tool do the thinking for the student?

Because like I said back in 2010,

A tool is just a tool! I can use a hammer to build a house and I can use the same hammer on a human skull. It’s not the tool, but how you use it that matters.”

Ultimately our focus needs to be on what we can and should use any tool for, and AI isn’t an exception.

I don’t ever ‘want to’ see ‘wanna’

Dear Siri,

I love speech to text. When I’m on the go I want to just speak into my phone and not bother typing. This is such a handy way to get words into a text or email with minimal effort. But this is the age of AI and it’s time to grow up and get a little more intelligent.

I don’t ever ‘want to’ say ‘wanna’.

“I don’t ever wanna see that word as a result of you dictating what I say.” (Like you just did.)

Listen to me Siri, I know my diction isn’t perfect. I know I don’t always enunciate the double ‘t’ in ‘want to’. But after I’ve gone to settings and created a text replacement shortcut from wanna to want to;

After I’ve corrected the text you’ve dictated from wanna to want to hundreds of times… Can you learn this simple request?

Don’t ever assume I want to say wanna!

Please.

Consciousness and AI

I have a theory about consciousness being on a spectrum. That itself isn’t anything new but I think the factors that play into consciousness are: Basic needs, computational ‘processing’, and idleness. Consciousness comes from having more processing time than needed to meet basic needs, along with the inability for processing (early thinking) to be idle, and so for lack of a better word desires are created.

Think of a very simple organism when all of its needs are met. This isn’t a real thought process I’m going to share but rather a meta look at this simple organism: “I have enough heat, light, and food, what should I do with myself?”

  • Seek better food
  • Move closer to the heat or light source
  • Try different food
  • Join another organism that can help me
  • Find efficiencies
  • Find easier ways to move
  • Hunt

At first, these are not conscious decisions, they are only a choice of simple processes. But eventually, the desires grow. Think decisions that start like, “If I store more energy I can survive longer in times of famine.” And evolve to more of a desire not just for survival but for pleasure (for lack of a better word): “I like this food more than other kinds and want more of it.” …All stemming from having idle processing/thinking time.

I don’t know when ‘the lights turn on‘, when an organism moves from running basic decisions of survival to wanting and desiring things, and being conscious? I believe consciousness is on a spectrum and it is idle processing/thinking time that eventually gets an organism to sentience. It’s sort of like the bottom of Maslow’s hierarchy pyramid must be met, (psychological and safety) AND there then needs to be extra, unnecessary processing time, idle time that the processor then uses for what I’m calling desires… interests beyond basic needs.

Our brains are answer-making machines. We ask a question and it answers, whether we want it to or not. If I say what does a purple elephant with yellow polkadots look like? You will inevitably think of what that looks like simply from reading the question. I think that is what happens at a very fundamental level of consciousness. When all needs are met the processors in the brain don’t suddenly stop and sit idle. Instead, the questions arise, “How do I get more food?”, “Where would be better for me to move to?” Eventually all needs are met, but the questions keep coming. At first based on simple desires, but more and more complex over generations and eons of time.

So why did I title this, ‘Consciousness and AI’? I think one of the missing ingredients in developing Artificial General (or Super) intelligence is that we are just giving AI’s tasks to complete and process at faster and faster times, and when the processing of these tasks are completed, the AI sits idle. An AI has no built in desire that an organic organism has to use that idle time to ask questions, to want something beyond completing the ‘basic needs’ tasks it is asked to do.

If we figure out a way to make AI curious, to have it desire to learn more, and to not let itself be idle, at that point it will be a very short path to AI being a lot smarter than us.

I’m currently listening to Annaka Harris’ audio book ‘LIGHTS ON: How Understanding Consciousness Helps Us Understand the Universe’ on Audible, and that’s inspiring a lot of my thinking. That said, this post is me rehashing an idea that I had back in December 2019, when I wrote, ‘What does in mean to be conscious?’… I go into this idea of idle time further in that post:

“…life requires consciousness, and it starts with the desire to reproduce. From there, consciousness coincidentally builds with an organism’s complexity and boredom, or idle processing time, when brains do not have to worry about basic survival. Our consciousness is created by the number of connections in our brains, and the amount of freedom we have to think beyond our basic survival.”

My conclusions in that post focused more on animal life, but listening to Annaka’s documentary of interviews with scientists I’m realizing that I really do think there is some level of consciousness right down to the simplest life forms. If it’s idle time and desires that bring about sentience, then figuring out how to make AI’s naturally curious will be the path to artificial general intelligence… Because they are already at a place where they have unused processing time, which is continuing to grow exponentially fast.

What it means to be literate?

Can you read? Can you do basic math? Is that enough?

The critical thinking required to make sense of the world today is ever increasing. We have a world leader using magical math to make a trade deficit calculation into a reciprocal tariff calculation, and claiming that this is, “Kind reciprocal, not full reciprocal.”

What? Help me make it make sense?

Meanwhile, I saw a video that someone created using AI. He uploaded a pdf article for two AI‘s to discuss, one of the AI’s was a version of himself, with his voice, and the other was a female at a desk. The only thing that suggested to me that the conversation was between two AI’s was some awkward hand gestures. Take those movements away, or make them a bit more natural/realistic and I would have no idea that I was watching an AI conversation.

Meanwhile, in Egypt, there are some wild claims about structures under the great pyramids, and while the evidence is unclear, I’ve seen many videos explaining these not-yet-proven structures. These claims include that they are a network of power sources connected to other structures around the world, and another theory claiming that aliens created them.

And speaking of aliens, wasn’t it just a few short months ago that we ‘discovered’ aliens living in our oceans? What ever happened to that story?

It’s becoming almost impossible to be informationally literate today. By the time you have time to seriously fact check something the story is already old, and there are new crazy claims that require your skeptical attention. What’s the source of this information? Where did they get their data from? What’s the bias of the news source? How is this data being manipulated? Who paid for the study? Is this a real quote? Is this video real, or CGI, or AI?

Who is fact checking the fact checkers? Meanwhile, here in Canada, a fact checker hired by one of our news stations was let go because trolls that don’t like their favourite political party being fact checked brought so much negative attention to her that the news station let her go.

What? Help me make it make sense?

The reality is that reading and writing and doing basic math is not enough to be functionally and informationally literate today. The critical thinking required to simply consume the information being thrown at us is overly demanding. I think the way forward for the short term is to find trusted sources and rely on them… and yet that’s the very thing that has seemed to get us into trouble. How many people get their news from just one or two biased sources? I’m literally now suggesting to find an echo chamber to sit in… hopefully you can find one that echoes facts, common sense, and some semblance of the truth.

AI text in images just keeps getting better

One of the biggest challenges with AI image generation is text. A new model, Ideogram 3.0 out of a Toronto startup, seems to have cracked the code. I wanted to try it out and so here were my two free prompts and their responses:

Prompt: Create an ad for a blog titled ‘Daily-Ink’ by David Truss.
The blog is about daily musings, education, technology, and the future, and the ad should look like a movie poster

Prompt: Create an ad for a blog titled ‘Daily-Ink’ by David Truss.
The byline is, “ Writing is my artistic expression. My keyboard is my brush. Words are my medium. My blog is my canvas. And committing to writing daily makes me feel like an artist.”

While the second, far more wordy prompt was less accurate, I can say that just 3 short months ago no AI image model would have come close to this. Now, this is coming out of a startup, not even one of the big players.

I didn’t even play with the styles and options, or suggest these in my prompts.

As someone who creates AI images almost daily, I can say that there has been a lot of frustration around trying to include text… but that now seems to be a short-lived complaint. We are on a very fast track to this being a non-issue across almost all tools.

Side note: The word that keeps coming to mind for me is convergence. That would be my word for 2025. Everything is coming together, images, text, voice, robotics, all moving us closer and closer to a world where ‘better’ happens almost daily.

Not emergence but convergence

My post yesterday, ‘Immediate Emergence – Are we ready for this?’ I said, “Think about how fast we are going to see the emergence of intelligent ‘beings’ when we combine the brightest Artificial Intelligence with robotics…” and continued that with, “Within the next decade, you’ll order a robot, have it delivered to you, and out of the box it will be smarter than you, stronger than you, and have more mobility and dexterity than you.”

On the technology front, a new study, ‘Measuring AI Ability to Complete Long Tasks’ proposes: “measuring AI performance in terms of the length of tasks AI agents can complete. We show that this metric has been consistently exponentially increasing over the past 6 years, with a doubling time of around 7 months. Extrapolating this trend predicts that, in under five years, we will see AI agents that can independently complete a large fraction of software tasks that currently take humans days or weeks.

More from the article:

…by looking at historical data, we see that the length of tasks that state-of-the-art models can complete (with 50% probability) has increased dramatically over the last 6 years.

If we plot this on a logarithmic scale, we can see that the length of tasks models can complete is well predicted by an exponential trend, with a doubling time of around 7 months.

And in conclusion:

If the trend of the past 6 years continues to the end of this decade, frontier AI systems will be capable of autonomously carrying out month-long projects. This would come with enormous stakes, both in terms of potential benefits and potential risks.

When I was reflecting on this yesterday, I was thinking about the emergence of new intelligent ‘beings’, and how quickly they will arrive. With information like this, plus the links to robotics improvements I shared, I’m feeling very confident that my prediction of super intelligent robots within the next decade is well within our reach.

But my focus was on these beings ‘emerging suddenly’. Now I’m realizing that we are already seeing dramatic improvements, but we aren’t suddenly going to see these brilliant robots. It’s going to be a fast but not a sudden transformation. We are going to see dumb-like-Siri models first, where we ask a request and it gives us related but useless follow up. For instance, the first time you say, “Get me a coffee,” to your robot butler Jeeves, you might get a bag of grounds delivered to you rather than a cup of coffee made the way you like it… without Jeeves asking you to clarify the task because you wanting a bag of coffee doesn’t make sense.

These relatively smart, yet still dumb AI robots are going to show up before the super intelligent ones do. So this isn’t really about a fast emergence, but rather it’s about convergence. It’s about robotics, AI intelligence, processing speed, and AI’s EQ (not just IQ) all advancing exponentially at the same time… With ‘benefits and potential risks.

Questions will start to arise as these technologies converge, “How much power do we want to give these super intelligent ‘beings’? Will they have access to all of our conversations in front of them? Will they have purchasing power, access to our email, the ability to make and change our plans for us without asking? Will they help us raise our kids?

Not easy questions to answer, and with the convergence of all these technologies at once, not a long time to answer these tough questions either.