Transparent heart icon with white outline and + sign.

Life's busy, read it when you're ready!

Create a free account to save articles for later, keep track of past articles you’ve read, and receive exclusive access to all RP resources.

White magnifying glass.

Search thousands of RP articles

Helping you think, speak, and act in Christ.

Open envelope icon with @ symbol

Get Articles Delivered!

Helping you think, speak, and act in Christ. delivered direct to your Inbox!





Red heart icon with + sign.
Assorted

How to catch Ice Age Man using digital reindeer

What would happen if we assumed ancient man was intelligent? One man did just that, and I haven’t been able to stop talking about it since. Which, incidentally, is a substantial reason to have 10 children, all of whom you homeschool. Captive audience. Professor John O’Shea is an archaeologist with the University of Michigan who likes to read. (It’s a good habit. Highly recommend.) One day, he was reading a book about subarctic reindeer* hunters and the stone structures they use to direct the animals. He was also studying a topographical map of the bottom of Lake Huron. The map included a feature known as Six Fathom Shoal, a continuous underwater ridge connecting southern Ontario to northeast Michigan. During the last days of the Ice Age, that ridge would have been above water, separating Lake Huron into two lakes. As John studied the map and read about the hunters, he realized that the ridge would have been a great place to hunt reindeer thousands of years ago. So he took a team and scanned the lake bottom. It wasn’t easy. The underwater ridge encompasses hundreds of square miles of lakebed, and just reaching the site required a 60-mile trip over unpredictable water. Expectations were low. Often the weather was too dangerous to even think about sailing, but on the rare occasions when they could get out of their port in Alpena, they used side scanning sonar to create a map of the lake bottom. One day, by “dumb luck” they stumbled upon something that maybe only John would have expected or recognized; a line of rocks. Stick with me here. On the leftside, of course You see, reindeer are essentially British. They like queues. Lines. Geometry, one might say, is their love language. To direct reindeer, a hunter can create a line on the ground, usually with stones – sometimes it amounts to a low fence This is called a “drive line.” When the reindeer encounter the line, even though they could easily jump over it, they will often walk alongside it wherever it leads. Not unlike Harold and the Purple Crayon. The line of rocks John O’Shea found on the bottom of the lake looked a lot like a drive lane. Further investigation by a tethered, unmanned underwater robot, and later, divers, revealed it was indeed a man-made drive lane, pointing towards a corral and a blind where the hunters could wait to kill the animals. An incredible, career-changing, history-making discovery. But it gets better. Next, John teamed up with Bob Reynolds from Wayne State University, and together they created a digital topographical map of Six Fathom Shoal during the Ice Age. They simulated rain, and the map filled with rivers, lakes, waterfalls, marshes and creeks. They enlisted botanists to look at the landscape and predict where different plants would have grown. And once the computer world was complete, they did the next sensible thing. They turned an entire herd of AI reindeer loose on the landscape. Initially, the AI reindeer had an annoying habit of walking off cliffs, gliding on stiff, pixelated legs to a swift digital death. The programming took a while to tweak, but as the weeks went on, the AI reindeer took on an increasingly lifelike set of behaviors. The herd was sent from one side of Six Fathom Shoal to the other, and patterns emerged as they learned which routes were best. Then John’ team asked modern reindeer hunters from Alaska and Canada to walk through the virtual landscape on computers, pointing out the best locations for blinds, campsites, and caches. They took intelligent human beings and asked them where they would hunt and then assumed that ancient man would hunt in the same places. Shocking. With this information in hand, they returned to inspect the sea floor. And that’s how they discovered a hunting structure they called “drop site 45.” Right in the location pinpointed as a choke site for the AI reindeer. This hunting site is the most sophisticated of its kind discovered to date. It includes a drive line, blinds, fire pits, and stone rings that indicate places where the hunters may have camped. They also found stone structures identical to those used by modern subarctic hunters to store caches of meat during the freezing winter months. Further investigation revealed stone tools, including one made with obsidian from Oregon. Yep. Oregon. Wagontire, Oregon, to be exact, over 2,000 miles away from the underwater ridge where it was found. Not bad for Ice Age hunters. “None of this matches the models we had about peoples in this region… you have to go back and be like, ‘All right, now we have this new data, what does that mean for what we thought about peoples that were living in the Great Lakes?’ You have to rewrite the story.” – Ashley Lemke, Professor at the University of Wisconsin Milwaukee Breaking down or evolving up? She’s right, of course. The models she’s talking about are based on the assumption that humans evolved from ape-like creatures, stumbling out of the dull fog of an animal existence to come to a shocking realization: “Fire… hot?” Of course we had never looked for evidence of Ice Age structures at Six Mile Shoal before, because the odds of such a creature building anything were slim. But what could we accomplish if we started with the assumption that man was created intelligent and capable, as the Bible teaches? Adam and Eve, coming straight from the hand of the Creator, had perfect genes. Those genes have only deteriorated a little more with each generation that passes. So those Ice Age hunters could well have been more intelligent than you or I. Computer models and sonar can recreate the Ice Age shorelines along the entire globe. What would happen if we looked at those maps and asked where we would build our cities, ports or villages? In some places, accumulated sediment has probably buried any Ice Age structures, but some remain. At the moment, our preferred method of finding these sites is dumb luck. That’s how we discovered a similar hunting drive line under the Baltic sea. The entire underwater world of Doggerland was also discovered by a series of accidents. What if we trusted the Biblical account enough to use less accidents and more “AI reindeer”? Shannon Skaer is an ancient history aficionado, homeschool mom, science nerd, and beekeeper. She lives in a small logging town in Oregon with her very own hero and their ten children. Her life could best be described as a love affair with beauty – which explains the number of babies – and her greatest joy is to know God, from whom all beauty comes. Find out more about her at www.shannonskaer.com. ***** *Reindeer is the fun name for Caribou. For this article, I refer to them as “reindeer” because I’m not a party pooper, and neither are you....

Red heart icon with + sign.
Culture Clashes

What will Canada look like in 2040?

Where some see gloom, God’s people can proceed in hope, knowing that God remains in control ***** “Canada to Become a Dystopian Nightmare, Households Will Flee: Gov Report.” A Christian friend shared this article with me, from Better Dwelling, a news outlet specifically focussed on Canada’s housing market. This friend wanted to make sure I was aware of a government report, “Future Lives: Social mobility in question,” that was published in January but which hasn’t been noticed and circulated until more recently. According to Better Dwelling’s Stephen Punwasi, the report from the federal government’s own think tank, Policy Horizons: “paint a grim picture resembling a dystopian mashup of a Charles Dickens’ novel meets Terminator. A Canada where wealth & the ability to own a home are determined at birth, hungry households hunt & fish for sustenance in cities, and moving down social classes is the norm. Welcome to Canada in 2040.” The message will undoubtedly resonate with many Canadians who have noticed that things have been changing quickly in the past five or ten years. Not so long ago, children could anticipate earning more, and spending more than their parents ever did. Now we may make more money, but it doesn’t go nearly as far. We can go to university, but a degree doesn’t mean much when it comes to getting a good job today. Young adults are finding it hard to imagine being able to buy even a modest home. Those that can find jobs are working more, and not having as many children. So, instead, to keep our population growing, we are bringing in millions of immigrants. But they need a place to stay too, which makes it even harder and more expensive for everyone to find a place to live. Then there is AI: that mesmerizing but creepy technology that is replacing many jobs and seems to be a lot smarter than most people using it today. Indeed, anticipating 2040 can be rather scary. But that is only true if we aren’t looking to the future with faith in our sovereign and loving God. PROVIDENTIAL GLASSES In Lord’s Day 10 of the Heidelberg Catechism, we confess that God upholds heaven and earth and all creatures so that “all things come to us not by chance but by His fatherly hand.” As a result, “with a view to the future we can have a firm confidence in our faithful God and Father that no creature shall separate us from his love.” I don’t know what is going to transpire in the next 15 years, but based on this confession of God’s providence, I believe a good case can be made that the changes we are experiencing aren’t something to fear but can be means through which God is gathering His Church and kingdom. What follows are some possibilities of the future, when looking through providential (in contrast to rose-colored) glasses. ARTIFICIAL INTELLIGENCE AI will have us seeking something authentic From my limited experience, the hype around AI is well-warranted. It is far “smarter” and more capable than I would have imagined. Even the over-used word “revolutionary” may not suffice. The advance of AI means we’re soon going to have a hard time telling whether what we read, see and hear is real or original: Did Emma really make that valedictorian speech that had us laughing and crying? Or was it the product of Chat GPT with a few tweaks to make it look authentic? Was that YouTube clip about the New York Yankees having a moment of silence after Charlie Kirk’s death real, or just AI-generated? How can we even know? Did the pastor actually write that sermon on Lord’s Day 10, or did he ask AI to make a sermon for him, pointing it to www.TheSeed.info to ensure that the result would line up with solid Reformed orthodoxy? Can I trust that the person calling me to ask for money is actually my son/grandson in trouble? He sounds just like him, but something just doesn’t seem right. This is just a small taste of AI’s impact and is legitimately concerning. God’s Word remains trustworthy But when we look through the glasses of God’s providence, something else becomes clear: in a world where it is very difficult to know what is true, solid, trustworthy, and real, the things that are will become all the more noticeable and meaningful. And what is more true, solid, trustworthy, and real than God and His Word? “The Lord is my rock, my fortress and my deliverer; my God is my rock, in whom I take refuge, my shield and the horn of my salvation, my stronghold” (Psalm 18:2). Just as many young men today are pushing against the woke culture they were catechized in, I won’t be surprised if many people begin recognizing the beauty of God’s Word for what it is: unchanging, reliable, proven, not open to opinion, and anything but artificial. In that dizzying sea of AI will stand the rock of God’s Word and the sure hope of the Gospel. That may explain in part why popular secular influencers like Joe Rogan, Piers Morgan, and Andrew Schultz are all asking Canadian Christian Wes Huff to come on their platforms to explain how Scripture is reliable. Who would ever have seen that coming this year? The Colson Center also shared news about a “quiet revival” across England and Wales: “the number of 16- to 24-year-olds in the U.K. attending church at least once a month jumped from 4% in 2018 to 16% in 2024.” To add to this, the most recent data from book sales revealed a 22 percent increase in Bible sales in the USA (compared with 1 percent for total book sales). And the “religious books” category saw the largest increase in publishing in 2024. For the past 75+ years, many in Western Civilization saw God’s Word as a relic from the past, that isn’t all that relevant. That seems to be changing. IMMIGRATION So many people With a plummeting fertility rate, Canada, along with most Western nations, relies on immigrants to keep our population and economies stable, let alone grow. In just two years, from 2022 to 2024, Canada’s population grew by 2,358,697 with about 98 percent of that increase due to incoming temporary and permanent residents. Such a large influx of new people over a short time can result in challenges. One has only to follow the news in places like the UK and Germany to see how difficult it is to provide leadership in a secular country in which many immigrants have little interest in upholding the social conventions and laws of that land. So many who can now hear But when we look through providential glasses, here too we see some amazing possibilities for the Church. My friend and his family were missionaries in a remote region of Africa, carrying God’s Word on foot to people immersed in paganism. But they came back to our small community in northern BC not long ago and realized that a lot has changed since they left for Africa. God has literally brought people from around the world to our own doorstep. This friend has decided to continue his mission work at home, reaching out to immigrants in our own community. At the same time, many “Canadians” who grew up in this nation have become as pagan, or more, than many of places where these immigrants are coming from. Millions of people in our own provinces aren’t familiar with the Gospel. The fields are ripe for the harvest, and they are next door! Ironically (providentially), God is sometimes using immigrants to challenge the trajectory that our society, and even some churches, have been on. For example, recent election results show that immigrants and minorities are swinging to the political right, favoring conservative parties federally and provincially. Closer to home, delegations representing different ethnic communities within the Christian Reformed Church urged their synod to adopt a biblical perspective of sexuality and were one of the forces leading to a shift in direction within that denomination. God isn’t looking to us to save Western civilization. Civilizations have risen and fallen many times. It is His kingdom that endures. And in His grace, God is bringing many to our land who are willing to “seek first His kingdom.” SOCIAL WELFARE SYSTEM The partial collapse of government The “Future Lives” government report mentioned at the beginning of this article made waves in part because it warned that in the dire near future “people may start to hunt, fish, and forage on public lands and waterways without reference to regulations. Small-scale agriculture could increase.” To add to this “governments may come to seem irrelevant if they cannot enforce basic regulations or if people increasingly rely on grass-roots solutions to meeting basic needs.” In other words, if Canadians are struggling under our socialized government welfare system, they may just start to take matters into their own hands and provide for themselves and their families. We saw a taste of this during Covid, as rural land became much more popular to own, and the public trust in government regulations was broken (in spite of daily assurances from public health officers). More freedom to be productive? I don’t think I need to work hard to convince most readers that there is a lot to this “warning” that may be a providential blessing. In a land where government regulation has stifled productivity and development for decades, many Canadians and Christians would welcome increased freedom and responsibility so that they can fulfil the cultural mandate that God gave humanity to “be fruitful and multiply, fill the earth and subdue it” (Genesis 1:28). That said, I think it is also the case that many Christians have become reliant on government income (i.e., monthly child care benefits, independent school funding, and even funding for their magazines) and may have a hard time adjusting if this were to decrease or terminate. Society will be greatly blessed when more responsibility is put on individuals, families, churches, and community organizations, rather than the state. But that is only the case if these people use their responsibility for the furtherance of God’s kingdom, and not their own. UNIVERSITIES Ivory towers may topple The government report also warned that “post-secondary education could be a stranded asset.” A stranded asset loses its value prematurely, as a result of a shifting market – think of someone who had a horse buggy factory as motor cars were taking over. University degrees might become like that? “People may look for alternative forms of training in new niches that appear to offer upward mobility. Non-traditional providers, including private firms, may outcompete traditional PSE players in attracting consumers.” What could rise from the rubble? As with the decline of the social welfare system, not many Christians will lament the breakdown of post-secondary education as it is represented in much of Canada today. Universities have been bastions of evil in our land, training generations of Canadians to undermine the Christian heritage that our civilization and country was founded on, and replace it with hedonism and, more recently, critical theory. There is little surprise that the “centers of higher learning” don’t help with “upward mobility” and might be outcompeted by private firms that seek to build, rather than tear down. The question is, what will be the worldviews of the private firms that are built? Will they, like Harvard and Princeton University were when they were founded, operate on a biblical and Reformed foundation? If so, unlike Harvard and Princeton, will they stay true to their mission? CANADA IN 2040 I understand why people aren’t optimistic about the future of Canada. Even in my own community, far removed from any urban centers and which, until recently, was known for being an idyllic place to raise a family, there are places that feel eerily similar to Vancouver’s Downtown Eastside. I see more homelessness, open drug use on the streets, and the need for 24/7 policing. My family isn’t even comfortable walking down parts of Main Street anymore. The same is true in communities across Canada today. This is the natural fruit of a secular worldview (ironically labelled “progressive”), and we can expect the trajectory to only continue as long as our country refuses to humble itself before the LORD. But God’s kingdom is above all of this and is advancing perfectly according to His plan. He put us in this time and place intentionally. He has a calling for us, right here and right now. We can leave the future in God’s hand, confident that He has the authority and power to guide all things. “If God is for us, who can be against us?” (Romans 8:31). There may be reformation or revival, like in the time of Josiah (2 Kings 22-23). Or God may have good plans for another civilization to take our place, as He has done to the Babylonian and Roman empires, along with so many others. Or He may usher us into glory today yet. God doesn’t burden us with the future. Our task is to focus on the present. We can use each day He has given us to build on the foundation of Jesus Christ, confident that “if the work that anyone has built on the foundation survives, he will receive a reward” (1 Cor. 3:14). As such, it doesn’t matter so much if Canada still exists in 2040. We know with certainty that God’s kingdom will endure....

Red heart icon with + sign.
Internet

How to use AI like a Christian boss

Imagine you’re the boss of your own company. After months of searching, you’ve just landed the most brilliant assistant in the country. He’s lightning-fast, top of his class in every subject, and available 24/7. He never sleeps, never complains, and never forgets a deadline. You can ask him for ideas, outlines, editing help – even technical research – and he’ll give you an answer in seconds. Best of all? He’s already sitting at your desk. His name is AI. Now here’s the catch: he’s not human and doesn’t share your values, or know right from wrong, and he always speaks with confidence even when he’s completely wrong. If you hand over your work to him, he might give you something that sounds smart, even impressive. But he might also serve up something misleading, shallow or just plain false. And because you’re the boss, it’s your name – and your integrity – on the line if anything goes wrong. Most of us aren’t bosses of our own companies…but every Christian has a calling to steward time, talents and resources. From the beginning, God made man to rule over creation (Gen. 1:26), and that includes ruling – rather than being ruled by – technology. There are valid reasons to approach AI with caution (as we’ve seen in previous Reformed Perspective articles like “Will AI Replace Reading?” and “Is AI Just Another Tool – or Something More?”). Even so, AI is here to stay, and it’s already reshaping the job market, communication, and everyday life for many people. One global management group says AI has the potential to be as transformative as the steam engine. AI is a tool unlike anything we’ve seen before, but at its core, that’s what it remains: a tool. And like any tool, it can be used for good or evil, depending on the people designing and directing it. Used wisely, AI can be an excellent assistant, capable of drafting hundreds of words in a short time. However, not all that AI produces is wise, relevant or true. That is why it’s important to think critically and test everything it says. Proverbs 14:15 reminds us, “The simple believes everything, but the prudent gives thought to his steps.” Just as it would be unwise to operate heavy machinery without proper training, Christians should not use AI without preparation and thoughtfulness. Quick answers can be tempting, but Proverbs 21:5 reminds us: “The plans of the diligent lead surely to abundance, but everyone who is hasty comes only to poverty.” Rushing to use AI without wisdom or careful review can lead to shallow or even dangerous results. I work as a project coordinator for a company that produces curriculum and I use AI almost daily in my work. It’s helped me draft content, edit writing, brainstorm ideas, and even develop Christian material (it does know a thing or two about Reformed theology). But I’ve also seen how quickly it can go off course. Used wisely, AI can be part of faithful stewardship. If you’re considering AI – or already using it – these seven principles can help you use it to the glory of God without compromising convictions or integrity. 1. Be the boss – not the bystander AI is here to assist, not lead. Think of it like a new apprentice: helpful, fast, and tireless – but not wise. AI can draft an article, summarize a report, or give you a list of ideas, but it doesn't know whether those ideas are any good. That’s your job. Use AI to boost your productivity, not replace your discernment. If you're an engineer or electrician, you know how this works already. An apprentice can be a huge help – they might prep materials, run calculations, or handle basic wiring to save you time. But when it's time to sign off on the plans or certify the work, it's your name that's on the line. If the apprentice makes a mistake and the building collapses or catches fire, you're the one held responsible. That’s why every detail needs to be carefully checked and approved by the licensed professional. AI is no different. It's an assistant – not the one who signs the final plans. 2. Think critically and watch for mistakes AI tools are designed to sound convincing – but convincing doesn’t always mean correct. Sometimes they generate information that looks polished but is actually shallow, misleading, or outright wrong. This is known as a “hallucination.” For example, a US lawyer who used AI for legal research is now facing his own court hearing after using false AI-generated information in court. The lawyer didn’t realize that several of the legal cases that AI had cited for him didn’t actually exist. He passed them along unchecked. I’ve experienced many hallucinations myself, such as when I asked AI to clarify a punctuation rule and it said one thing in the rule and presented the opposite in the example it gave. That’s why you can’t just copy AI’s response and hit “send.” You need to review the results carefully. If you’re using AI to explore a topic you don’t know well, make sure you double-check the facts, confirm the logic, and – if possible – ask someone with more experience to give it a second look. AI has saved me significant time researching unfamiliar topics, but before finalizing anything, I verify the sources or have someone with expertise review it. Ideally, you should have at least some grasp of what good work looks like in the area you’re using AI for. If not, treat the AI’s output as a starting point, not a finished product. Use it to learn, refine, and check your thinking… but don’t assume it’s right. 3. Train AI like an apprentice What AI gives you after your first prompt is often just a rough draft. The result might be serviceable, but it’s rarely great unless the task is very simple. After all, apprentices need training. Here are some key tips for getting better results from AI: • Tell AI the role it should take on (math teacher, history professor, writer, business expert, travel agent, event planner, etc.). • Outline as many details as possible – task, tone, purpose, websites it can research, intended audience, length. • Provide examples. • Ask AI what questions it has for you. • After reviewing the output, point out how AI can improve the results. • Do a few edits of your own and let AI know what you did for future reference. For example, a first prompt for writing could look like this – you would tell ChatGPT: “Assume the role of an expert copywriter, familiar with Reformed theology as taught by John Calvin and R.C. Sproul. You are deeply familiar with the Heidelberg Catechism, Canons of Dort and Belgic Confession. Your job is to write articles for Reformed Perspective magazine. Here are some writing guidelines for this magazine… “I'd like to write an article for this magazine titled ‘How to Use AI Like a Christian Boss.’ In it, I'd like to compare AI to a smart apprentice. You can use what they come up with, and it might be very good, but before an engineer or architect can put a stamp on it, they have to go through every detail and make sure that they stand behind it… “Write this article in prose, but structured with numbered points for how to use AI like a Christian boss. Start with an outline. But before you do that, what questions do you have for me?” The first prompt is just the beginning. After that, a “Christian boss” process would look like: • Answering AI’s questions. • Instructing it to draft the outline for the article. • Carefully reviewing the outline, making some refinements yourself and/or telling AI to make refinements. • Instructing AI to draft the article. • Doing a detailed review and editing. The edits can be done on your own or by prompting AI (see the next point for some tips on that). I usually do a mix of both personal and AI-prompted edits. With refinement, specific instructions, and key edits, working with AI can turn a mediocre first result into a solid piece of writing. Note that because getting an excellent AI response often requires multiple rounds of prompting and giving feedback, it’s not always faster to use it for a task that you only need to do once. 4. Use AI like a creative sidekick AI can be a great help when you need a creative boost. You might ask it to: • Rewrite a sentence five different ways so you can choose the best version. • Give practical examples to strengthen a concept you’re trying to explain. • Help you generate ideas for starting or improving a project. • Reword a section of text for clarity, flow, or tone. • Give feedback and ideas for improvement on something you’ve written. • Ask you questions to help you think about something in new or deeper ways. Sometimes AI comes up with something surprisingly helpful. Other times it completely misses the mark. The key is to use it as a creative partner, not a crutch. Be prepared to write things yourself if AI just doesn’t give you what you need. (Yes, AI can have an off-day too.) 5. Protect your privacy – and your mind AI tools aren’t generally private. What you type may be stored or used to train future models, depending on the platform. That means anything personal or sensitive might not stay confidential. So be cautious. Don’t share anything you wouldn’t want repeated or misunderstood. Because it can mirror your tone and affirm your ideas, AI can start to feel personal, to the point that some people have started treating it like a friend, therapist or romantic partner. However, hearing exactly what you want from a machine can pull you away from real, God-given relationships. God calls us to grow in community, where we can be encouraged in our faith and held accountable when we wander. Proverbs 27:17 says, “Iron sharpens iron, and one man sharpens another” (ESV). AI can echo your own voice, but it will never call you to repentance, speak truth in love, or walk alongside you in genuine discipleship. 6. Train for discernment before you use it Just as students need to understand what 2 + 2 means before using a calculator, Christians need foundational knowledge before turning to AI. Skills like reading, writing, theology, math, and logic help us recognize when AI is inaccurate, shallow, or biased. AI is trained on massive amounts of data, and although it can recite the Heidelberg Catechism, most of its data likely didn’t come from a Christian perspective. One English teacher shared that when her students used an AI tool to give feedback on their writing, it consistently flagged Christian content as “too one-sided.” Romans 12:2 warns: “Do not be conformed to this world, but be transformed by the renewal of your mind, that by testing you may discern what is the will of God, what is good and acceptable and perfect.” God calls us to use discernment. Without a strong grounding in Scripture, truth, and general knowledge, we won’t have the tools to spot harmful ideas or use AI wisely. 7. Don’t let AI replace real mentors AI can be a useful support for learning, without replacing skill development. It can offer feedback, generate ideas or ask helpful questions to deepen your thinking. But it should never replace the guidance of real people. For Christians, learning isn’t just about improving skills or producing results. It’s about growing in wisdom and character. AI can’t be trusted to help you think biblically, challenge you in love or walk you through real-life decisions. That happens best in relationships. God commands, “Train up a child in the way he should go; even when he is old he will not depart from it” (Prov. 22:6). That is not a task we should trust to AI. If young people are going to use AI, they need a strong biblical foundation and wise guidance. Without that, AI becomes a shortcut rather than a tool – and we can’t shortcut godly wisdom and discernment. Conclusion Used wisely, AI can strengthen our work and spark new ideas. But as Christians, we don’t just care about what works, we care about what honors God. “The fear of the Lord is the beginning of wisdom” (Prov. 9:10). Faithful stewardship isn’t measured by cleverness or creativity, but by our trust in Christ and obedience to His Word. AI reflects the priorities of the person using it. So let’s bring ours under the lordship of Christ. Take responsibility. Stay alert. “Whatever you do, work heartily, as for the Lord and not for men” (Col. 3:23). Whether you use AI or not, let this be your aim: to honor God in all things, rule over creation, and never let created tools rule over you. P.S. In case you’re wondering, I did use AI like a Christian boss to write this article. See the short article below. ***** WHOSE SPEECHES WERE THEY? A quick conversation on having AI, and others, writing for us JON DYKSTRA: The one question I know readers will be asking, so let’s give them an answer, is, approximately what percentage of the article is AI written? Or is that even something you can put a percentage on? VALERIE VANDENBERG: That percentage question is a tough one. My process with AI involves a lot of back and forth. To give some more details, by the time AI drafted the article, I had already given it about 1,500 words of my own instructions (including the ideas I had for the article and answering AI's questions for me) in addition to giving it your writing guidelines from the Reformed Perspective website. JD: Can you get into the process just a bit more? VV: My first prompt included my ideas for the article and detailed instructions for my vision for it. Then I had AI draft an outline, which I adjusted until I was satisfied. After that I instructed AI to draft the article itself, which was followed by detailed editing (sometimes done by me, sometimes prompting AI to edit a section, and usually a mix of both). I often have AI write things a few different ways so I can glean the best ideas from the list. Or I just write it myself if I think my idea is better than what AI suggested. JD: This is something Reformed Perspective staff have been wrestling with, trying to think through the extent or limits we’d want to use ChatGPT or other AI. To this point I haven’t used it to generate text, but that’s been more a hesitancy – I’d like to listen in on the debate some more before coming to a firm conclusion – than any specific principled objections. Your article is a part of that debate, and I’m grateful for it. My initial take is that using AI to generate text, as in this article, is akin to a president using a speech writer. The writer crafts the words, but the president sets the direction, and adds in his own tweaks and orders rewrites, such that at the end he will so completely own these words, that we will fully attribute this speech to him, and not the writer. VV: Yes. The key Christian boss part of the process is that by the time I hit send to submit the article, every aspect of it (tone, content, structure, wording, etc.) is something I can stand behind, and as good as I could do alone or ideally better....

Red heart icon with + sign.
Internet

Will AI replace reading?

Empty libraries and human-less humans ***** Many new technologies are sold with the promise of freeing people from menial tasks. Dishwashers, dryers, tractors, and word processors are just a few of the many inventions that have made life easier, reducing the amount of backbreaking labor involved in necessary chores and leaving users more time for things worthwhile, like learning, creating, and enjoying relationships. But what happens when technology promises to “free” us from even those worthwhile activities? That’s one of the many questions we face in the age of artificial intelligence. For example, entrepreneur and “Shark Tank” judge Davie Fogarty recently told his 40,000 followers on X that “(r)eading books is now a waste of time. AI reasoning models can distill key insights and tell you exactly how to implement them based on everything they know about you.” Can reading really be outsourced to AI? Should it be? Is this a post-schooling version of the new epidemic of AI-based cheating where students have chatbots do their research and compose their assignments? Is the study and reflection on ideas now as much of an historical anachronism as plowing a field by hand? The process is the point The belief that reading and writing should be delegated to AI betrays a confusion not only about what technology is for but, even more, what we are for. Also, it lands us in some dark places, philosophically and spiritually. Author and classics professor Spencer Klavan wrote on X that many students who outsource the slow work of reading and writing soon find themselves wondering what the point of life is. After all, why go on if humans are obsolete, and chatbots can perform every task better in a fraction of the time? All new technologies require humans to wrestle again with what it means to be human. Human work is vital, not only because it is a way in which humans love and serve our neighbors, but because it is a fulfillment of the creation mandate. It’s true the Fall has turned much of our work into toil, and so any technology that alleviates futile, dangerous, and pointless work is a blessing. However, for some of our work, like reading, the process is the point. Not all activities can be measured in the narrow, utilitarian way that Fogarty and other over-eager fans of AI claim. Assuming AI can “distill” a work accurately, or that we need to be made “free” of sitting with an author, following an argument, or experiencing a narrative reduces truth, goodness, and beauty to mere data. To optimize or automate reading is simply not to read. It is like asking AI to free us from eating a delicious meal or taking a walk in the park with our kids. Some things cannot be optimized or outsourced, because they are irreducibly embodied, conscious, and human. Required to read and wrestle The best case-in-point is the Bible. God could have revealed what He wanted us to know in a bulleted list of “distilled” theological “insights” or moral pronouncements. Instead, He gave us a library of stories, proverbs, epistles, history, and authors, writing diverse types of literature over centuries, all of which comprise Holy Scripture. Part of what makes the Bible such a gift is the work and the humility God requires of us as we wrestle with It. Describing the slow and divinely blessed act of reading the Bible as a “waste of time” is a failure to grasp what it is, why it was given to us, and why we, as creatures, need it. And this is also true of many other books. To “free” us from this wonderfully inefficient process is to free us of our humanity. It is asking to be liberated from the nature given us by the God whose image we bear. To paraphrase the Psalmist, this will be how humans in this technocratic age become like our artificially intelligent idols. Novelist and songwriter Joseph Fasano wrote a poem entitled “For a Student Who Used AI to Write a Paper”: I know your days are precious on this earth. But what are you trying to be free of? The living? The miraculous task of it? Love is for the ones who love the work. The most important part of that work, in fact, reflects what it means to be made in the likeness of God instead of a computer. For more resources to live like a Christian in this cultural moment, go to Breakpoint.org. This is reprinted with permission from the Colson Center....

Red heart icon with + sign.
Internet

3 things we need to tell our kids about ChatGPT

ChatGPT and the many other new AI text generators might strike parents as problematic, since kids are now able to turn to this tool and, with just a few prompts, churn out their homework. Need a 600-word essay on the biblical perspective on why abortion is wrong? ChatGPT can output it in less time than it’ll take most students to type the request. Isn’t that cheating? Yes, if the teacher wanted students to write it on their own. But it also isn’t hard to imagine how teachers could also incorporate AI tools into lessons on not only writing, but editing and analysis. How good is the AI essay? Where is it weak, and what might it be missing? Did the opening grab you? Would its argument be more powerful as a dialogue? What other prompts could we use to tighten it up?  AI possibilities are enormous but yes, ChatGPT does also open up new temptations for kids to shirk the work they are supposed to be doing on their own. That means that we, as parents, are going to have to remind and reinforce to our kids a few important points: 1) God loves a hard-won C. Or to put it another way, God doesn’t care about your marks: if you get a low grade but tried hard, great, but He hates an ill-gotten A. And your parents think the very same. 2) Cheating hurts you (Prov. 10:2). A basketball player might be able to build a robot that shoots better than he does, but it isn’t going to help him learn how to shoot. If an assignment is intended to help a student learn to write, getting someone or something to write it cheats the student out of what they could have learned. Cheating is also a matter of character – if you’ll cheat on something as little as an essay, what kind of person are you becoming? You do become what you do. 3) Knowing how to write remains an important skill even in the era of AI, because of all the skills a student has to learn to be able to write like: research, organizing thoughts, and learning to discern truth from error and stronger points from weaker ones. That’ll help you write an essay, but also choose a career, and even assess who you might want to marry....

Red heart icon with + sign.
News

Lawyer learns she should double-check ChatGPT’s work

Last month, a BC lawyer was caught submitting an AI-generated legal application to a family court. Chong Ke was representing a father asking that his children travel to China so he could have parenting time with them there. Ke used ChatGPT, a text-based generative Artificial Intelligence (AI), to write the application, which cited two fake cases. ChatGPT does not provide links to its sources which makes its fabricated content hard to spot. To the AI’s “success,” the case was won, finding it was in the best interest of the client's children to visit their father. Although the fake cases were put into the application, they were never presented in the hearing, so the courts kept the initial ruling. Shortly after the ruling, the lawyers for the mother tried to track down the cases referenced, even asking for copies from the opposition. After not receiving any copies the team hired a researcher to find the cases. This matter was then taken to court where Ke apologized for her actions. "I acknowledge that I should have been aware of the dangers of relying on Al-generated resources, and been more diligent and careful in preparing the materials for this application. I wish to apologize again to the court and to opposing counsel for my error.” The process of preparing legal documents can be arduous, and although AI technology might be a tool to help with this, we should not be naive about its capabilities. Ke was ordered to pay the costs associated with the time and resources the other counsel used to discover the fake cases. In attempts to make court applications easier, the lawyer ended up doing the opposite, deceiving others in a scramble to seek truth in the application. In the court case against Ke, the judge reminds us that AI is not a replacement for humankind. “As this case has unfortunately made clear, generative AI is still no substitute for the professional expertise that the justice system requires of lawyers."...

Red heart icon with + sign.
News

Saturday Selections – Feb. 17, 2024

Failing at failure? This could be a great way to bring up an important conversation with our kids. The fear of failure stops many children from trying new and hard things. God gave them talents, and if they are going to develop those skills (and not bury them - Matthew 25:14-30) our children will need to push their limits. That might mean running so hard they start tripping, stumbling, and even eating some dirt. They need to know there is such a thing as God-glorifying failure... but not God-glorifying cowardice. Marriage makes men better Though he doesn't acknowledge God, this evolutionist has discovered that God's plan for the family – marriage – works best, reining in men's antisocial behaviors. There is a really great metaphor here, too, about our rational self being a rider trying to control an elephant that represents our impulses. Those impulses – or, in Christian terms, sinful desires – can be easier or harder to control depending on where we take them. So, for example, a rider troubled by alcohol shouldn't "take his elephant" into a bar. To extend this to the family realm, a dad or mom trying to deal with a kid that pushes their buttons shouldn't indulge in watching the late show – they'll have to put their elephant to bed early each night to better enable them to be calm and controlled. The Swiss Family Robinson: a return to the classics Jonathon Van Maren makes a plug for this great book. Parents, if you get your kids a copy, be sure you get an attractive one: a great cover, good font, and a few illustrations thrown in here and there can really help by making a classic so much more appealing. G.K. Chesterton on AI If you want to understand AI, then who better to ask than someone born two hundred years before it was invented? This is a good one! And for an in-depth dive check out what's on offer at Creation.com. The most bizarre experience of my life When theologian E. Calvin Beisner was invited to do an interview about climate change, he never expected to be interviewed by a clown. Who knows how this will end, but some people certainly are desperate to try to make Christians look bad. "He Gets Us" takes a big "L" in the Super Bowl Among the Super Bowl commercials last week were two to promote the "He Gets Us" evangelistic campaign. Many Christians have defended them as "pre-evangelism" – sure, they didn't hint at the Good News, but they did tell people that Jesus gets them. That's a start, right? There's some truth to that – they had a minute, and if you had just a minute would you be able to present the whole of the Gospel to someone? To say it another way, an incomplete message isn't wrong... it just needs more. But the problem comes when we say the easy part and stop. And a lot of "churches" are content with just telling folks that Jesus gets them. But they don't want to offend anyone. Telling people they need to repent? Jesus as Saviour is offensive indeed. Now someone has shown how it is possible, in just a minute, to tell people quite a lot about Jesus and their need for a Saviour. And Ray Comfort gives his own version here. ...

Red heart icon with + sign.
Internet

Demystifying ChatGPT

Lots of potential, but the output is only as valuable as the input ***** About a year ago, the research firm OpenAI made a version of its text-generation tools available for free use from its website. The chatbot, called ChatGPT, accepts a prompt from the user, such as a question or a request for a piece of writing, and responds with a seemingly original composition. If you have experimented with this tool, you may be impressed with its ability to produce natural-sounding English paragraphs – or perhaps you find it eerie and wonder what changes tools like this will bring. How large language models work To demystify what language tools like ChatGPT are doing, here’s a game you can try at home. To play, you’ll need to select a book and pick a letter of the alphabet. I’ll play along while I write. A) One letter I have a copy of Pilgrim’s Progress handy, and I choose the letter m. Open your book at random and find the first word on that page that contains your letter. For me, the word is mind. In the word that you found, what letter occurs immediately after the letter you were looking for, just as i comes next in mind ? That letter is your new letter. Now flip to a new page and repeat. The first i that I see is in the word it, so that means t is my next letter. Do it again. On the next page that I turn to I see t in the word delectable, so a is my next letter. So far my letters have been m-i-t-a. On my next turn, I find a in about and take the letter b. I then find b in Beelzebub and take the letter e. Finally, I find e in the word be, and since it’s followed by a space rather than another letter, this ends the game. Looking back, the letters of the game make the word mitabe. That’s not a real word, of course, but it’s not completely random either. The letter-level statistics of the English language led this process to make something that looks more English-like than we are likely to have gotten by grabbing six tiles out of a bag of Bananagrams. B) Two letters This game becomes more interesting if we level it up. Instead of looking for the first occurrence of our most recent letter, what if we searched for the first time our last two letters occur together? If my last two letters were be, then I would need to a look for a word like become and take c as my next letter. The effect is that we are using a larger context as we string out our sequence of letters. This version of the game is tedious to do by hand, but we can automate it using a computer. Drawing from the text of Pilgrim’s Progress and using two letters of context, an algorithm replicating this game generates the text: SETTERE HEY IST ING TO WAS NOR HOUT SAY SUPOSTIANY If we sift through the gibberish, we find five real English words – and the rest are at least pronounceable. C) Four letters If we try again, but this time with a context of four letters, we get: WOULD BUT THIS ONE COUNTRY AND SMOKE TO HAD KEPT Although this is nonsense, all the words are real, and shorter strings of words could make sense together. D) Nine letters Finally, if we level up the game to use nine characters of context, sentences begin to take shape: SO CHRISTIAN AND HOPEFUL I PERCEIVE IS AGAINST A WALL Grammar aside, this is remarkably close to real English. When it gets bigger still To get a layman’s understanding of how technology like ChatGPT works, imagine a great leveling-up of this game. Instead of being played on one laptop, it is run on giant server farms of leading tech companies. Instead of using Pilgrim’s Progress as a source text, it is trained on humankind’s entire digitized corpus. Instead of using a few letters of context, it uses a sophisticated model of human language. And instead of using a starting letter, the process is induced by a question or request from a user. When you enter a prompt into ChatGPT, you are really asking a statistical question: Based on information gathered from innumerable digital texts, what is the most likely response to this prompt? I tried giving ChatGPT the prompt: “Versify the first question and answer of the Heidelberg Catechism into common meter so that it could be sung by a congregation.” It responded with: In whom is all my trust and joy, My comfort and salvation? In Jesus Christ, the sinless One, The source of my redemption. My only comfort here on Earth, My only hope in heaven, Is that I am not my own, but His, For all my sins are forgiven. The meter is forced, but one has to admit that many would-be hymn writers have done worse. Benefits and risks of text-generators ChatGPT and similar tools that generate text or images are one example of advances in technology that use algorithms built from data rather than being human-handcrafted. You probably have used some of them: data-generated algorithms can identify faces in photographs, control autonomous vehicles, make medical diagnoses, and detect fraudulent transactions. One effect of this trend is that the technologies become more difficult to understand, even for experts, since the tools are often shaped by deep patterns in the training data that are beyond human perception. They exude something of a magical quality, especially when they are presented with evocative terms like artificial intelligence. Yet it is important for Christians not to attribute anything magical to unfamiliar technologies. Even without precise expertise in trending technology, we still can develop an informed awareness of the benefits and risks. 1) Output only as good as the input For one, technology generated from data is only as good as the data it is generated from. A tool like ChatGPT reflects the attributes of the texts that it is trained on. I gave ChatGPT the prompt, “Explain how the Auburn Affirmation affected the career of J. Gresham Machen” . It responded with a page and a half of text that got the basic facts right and read like an answer on an essay test. But what stood out to me was the uniformly positive terms it used when referring to Machen. He was a “staunch defender of conservative, orthodox Christianity,” to which he had “unwavering commitment,” making a “courageous stand against the Auburn Affirmation” and founding the OPC, “where he could continue to champion conservative Reformed theology.” No doubt readers will be sympathetic to this portrayal. But it is worth asking why ChatGPT would give Machen heroic verbs and adjectives while describing the proponents of the Auburn Affirmation in dry, factual terms. My speculation is that people interested enough in Machen to write about him have tended to be his admirers, and so ChatGPT is imitating the dominant sentiment in material about Machen. But does this pass the shoe-on-the-other-foot test? What would ChatGPT produce if we gave it a prompt for which most of its source material was written by enemies of the gospel? 2) Truth isn’t a particular priority Furthermore, the fact that these tools produce the most probable response to a prompt, based on the data they are trained on, means that truth itself is not a particular priority. I once asked ChatGPT who was the most famous person to play against the eighteenth-century chess-playing automaton known as the Turk (actually an elaborate hoax). The response mentioned Napoleon Bonaparte and Benjamin Franklin – each of whom did face off against the Turk at one point – but it went on to claim that “one of the most famous opponents of the Turk was the Austrian composer Wolfgang Amadeus Mozart,” describing their chess match in dramatic detail. This appears to have been synthesized whole cloth – I can’t trace down any verification that Mozart ever encountered the Turk. One of my colleagues at Wheaton College, where I teach, described how he caught a student passing off a ChatGPT-generated term paper as his own work. The paper was more-or-less B+ quality, and its bibliography had respectable-looking citations. As it turned out, though, ChatGPT had made the citations up: the claimed authors were real scholars and the journals were real, but the articles themselves didn’t exist. This is what we can expect when the most probable response is not the most truthful response. 3) Temptation to students This serves to highlight the most obvious risk induced by ChatGPT and similar tools: it provides a new and perfectly convenient way to plagiarize. A student – or a professional writer, for that matter – can whip out a paper in seconds by giving a prompt to a free online tool. At present, teachers can use services that detect AI-origin of text (itself an application of data-driven algorithms), but we can expect that successive generations of text-producing tools will prove to be better at eluding detection. One can imagine an arms race ensuing between text generators and generated-text detectors, with teachers never being able to trust their students again. And what of the rights of the original authors whose work is used to train these tools? 4) Do away with drudgery On the other hand, it is all too easy to bemoan the potential harms in a new technology and to overlook how it can be used for good. Tools like ChatGPT can be used not only as a cheat for writing, but also as writing aides: You can feed it a paragraph you have written and ask it to polish it up – make it more formal, or less formal, or more succinct, or with a more varied vocabulary. It can act, in a way, as a smart thesaurus. Another colleague of mine told of a student who defended the use of ChatGPT in writing a paper as a natural progression from tools that are already accepted: If we use spellcheck to eliminate spelling errors and Grammarly to fix syntax, then isn’t it only wise to use the latest tools to improve our rhetoric as well? One may quibble with this student’s logic, but technological changes throughout history have automated menial tasks, allowing humans to focus on things that are more meaningful. After all, some writing tasks are so much drudgery. You do not want your pastor using ChatGPT to write a sermon, but perhaps you wouldn’t begrudge a businessperson using an automated writing tool that can turn sales data into a quarterly report. A Christian response Text-generation tools are part of a suite of data-derived technologies that have gotten much media attention and that seem to have the power to change society as much as the internet and smartphones have done, if not more. How should we live in light of advancements in science and technology? There are many ways Christians should not respond: we shouldn’t idolize technology or make a false gospel out of it – we should not share the world’s fascination for the next new thing or hope for scientific deliverance from life’s problems. But neither should we approach it with fear or regard it with superstition. I also would argue that it’s irresponsible to ignore it. First, to any extent that we use tools like ChatGPT, we should see to it that we work with complete integrity. This is in the spirit of Paul’s exhortation to Timothy that in his ministry of the Word he should be “a worker who has no need to be ashamed” (2 Tim. 2:15) – whatever our calling, our work should be worthy of approval. Students at any level of schooling should note their school’s and teachers’ policy on writing assistants and follow them scrupulously. The Christian school that my children attend has ruled that no use of ChatGPT is appropriate for any of their schoolwork, and, with admirable consistency, they have banned the use of Grammarly as well. Some of my colleagues have experimented with allowing limited use of ChatGPT as a writing aide in college courses, but with clearly-defined boundaries, including that students must cite to what extent ChatGPT was used. For any professional or personal use of these tools as writing aides, we should ask ourselves whether we are presenting our work honestly. Second, we should exercise a healthy skepticism toward material generated by tools like this. One use of ChatGPT is as a replacement for a search engine when looking up information. You can Google “how to keep wasps out of my hummingbird feeder” and receive a list of websites, each website giving advice about bird feeder maintenance; alternately, you can ask the same question to ChatGPT and receive a single, succinct summary of tips that it has synthesized from various sources. There is a pernicious draw to attributing authority to information delivered to us by computers. We should remember, though, that ChatGPT’s answers are only as good as the fallible human-produced text it is trained on. Moreover, for the sake of loving our neighbors, we should bear in mind that tools trained on data will reflect, if not amplify, the biases and prejudices of its input. Finally, if we use algorithms to manipulate text, we must treat the holy things of God as holy. In the experiment at the beginning of this article, I sampled letters from the text of Pilgrim’s Progress. When I do a similar experiment in one of the courses I teach, some students are curious what would happen if we sample from a book of the Bible. But I believe that is not a respectful use of God’s Word. Earlier I showed the result of asking ChatGPT to versify part of the Heidelberg Catechism, but I certainly do not advocate using AI-generated texts in congregational singing. One feature of ChatGPT is that it can imitate the style of an author or genre. You could ask ChatGPT to write a story in the style of the Bible, but don’t – that would be blasphemous. I find the Westminster Larger Catechism’s words about the sixth commandment – that it requires “a sober use of meat, drink, physick, sleep, labour, and recreations” – to be applicable to many other areas. Let’s pray for wisdom to discern the sober use of humanity’s tools in every age. The author is a member of Bethel Presbyterian in Wheaton, Illinois, and professor of computer science at Wheaton College. This is reprinted with permission of the OPC denomination magazine New Horizons (OPC.org/nh.html) where it first appeared in their January 2024 issue....

Red heart icon with + sign.
Internet

Is AI just another tool, or something else?

We must ask how the technology fits into the Creation story lest it takes away from meaningful work and connection. **** It’s not uncommon to hear artificial intelligence (AI) described as a new “tool” that extends and expands our technological capabilities. Already there are thousands of ways people are utilizing artificial intelligence. All tools help accomplish a task more easily or efficiently. Some tools, however, have the potential to change the task at a fundamental level. This is among the challenges presented by AI. If in the end it is not clear what AI is helping us to achieve more efficiently, this emerging technology will be easily abused. AI’s potential impact on education is a prime example. Since the days of Socrates, the goal of education was not only for students to gain knowledge but also the wisdom and experience to use that knowledge well. Whether the class texts appeared on scrolls or screens mattered little. Learning remained the goal, regardless of the tools used. In a recent article at The Hill, English professor Mark Massaro described a “wave” of chatbot cheating now making it nearly impossible to grade assignments or to know whether students even complete them. He has received essays written entirely by AI, complete with fake citations and statistics but meticulously formatted to appear legitimate. In addition to hurting the dishonest students who aren’t learning anything, attempts to flag AI-generated assignments, a process often powered by AI, have the potential to yield false positives that bring honest students under suspicion. Some professors are attempting to make peace with the technology, encouraging students to use AI-generated “scaffolding” to construct their essays. However, this is kind of like legalizing drugs: there’s little evidence it will cut down on abuse. Consider also the recent flood of fake news produced by AI. In an article in The Washington Post, Pranshu Verma reported that “since May, websites hosting AI-created false articles have increased by more than 1,000 percent.” According to one AI researcher, “Some of these sites are generating hundreds if not thousands of articles a day. … This is why we call it the next great misinformation superspreader.” Sometimes, this faux journalism appears among otherwise legitimate articles. Often, the technology is used by publications to cut corners and feed the content machine. However, it can have sinister consequences. A recent AI-generated story alleged that Israeli prime minister Benjamin Netanyahu’s psychiatrist had committed suicide. The fact that this psychiatrist never existed didn’t stop the story from circulating on TV, news sites, and social media in several languages. When confronted, the owners of the site said they republished a story that was “satire,” but the incident demonstrates that the volume of this kind of fake content would be nearly impossible to police. Of course, there’s no sense in trying to put the AI genie back in a bottle. For better or worse, the technology is here to stay. We must develop an ability to evaluate its legitimate uses from its illegitimate uses. In other words, we must know what AI is for, before experimenting with what it can do. That will require first knowing what human beings are for. For example, Genesis is clear (and research confirms) that human beings were made to work. After the fall, toil “by the sweat of your brow” is a part of work. The best human inventions throughout history are the tools that reduce needless toil, blunt the effects of the curse, and restore some dignity to those who work. We should ask whether a given application of AI helps achieve worthy human goals – for instance, teaching students or accurately reporting news – or if it offers shady shortcuts and clickbait instead. Does it restore dignity to human work, or will it leave us like the squashy passengers of the ship in Pixar’s Wall-E – coddled, fed, entertained, and utterly useless? Perhaps most importantly, we must govern what AI is doing to our relationships. Already, our most impressive human inventions – such as the printing press, the telephone, and the internet – facilitated more rapid and accurate human communication, but they also left us more isolated and disconnected from those closest to us. Obviously, artificial intelligence carries an even greater capacity to replace human communication and relationships (for example, chatbots and AI girlfriends). In a sense, the most important questions as we enter the age of AI are not new. We must ask, what are humans for? And, how can we love one another well? These questions won’t easily untangle every ethical dilemma, but they can help distinguish between tools designed to fulfill the creation mandate and technologies designed to rewrite it. ***** This column was first published to Breakpoint.org on January 8, 2024, and is reprinted with their gracious permission. We're sharing it because it's a good article on an important topic. But we have another reason. We wanted to give RP readers this sample of Breakpoint's Daily Commentaries to, hopefully, pique your interest. Breakpoint has an American focus and is not specifically Reformed (though some writers are), so we differ in some notable respects: they are anti-evolution and RP is specifically 6-day creationists; and while we'll highlight problems with the Pope both when he is acting Roman Catholic and when he is not, they'll stick to the latter. So, as with everything, there is a need to read with discernment. But when it comes to the hottest cultural battles of our day – sexuality, gender, the unborn, and God's sovereignty over "every square inch" of creation – they get it right, consistently, and they are timely, often replying to events that happened just the day before. That's why Breakpoint articles have been featured in our Saturday 6 column for years now. If this article did grab your interest, then you'll want to sign up here to get Breakpoint sent right to your email inbox each day....