Cursor

AI Policy

Embracing the Elephant in the Algorithm

AI has entered the chat, and we’ve given it a desk at 3 Sided Cube HQ. We’re not blind adopters or fearful sceptics. We’re curious humans using AI to elevate ideas, automate the boring bits, and leave more time for the stuff that actually makes a difference (like changing millions of lives for the better)

Our AI North Star

💚 Automate with intention - AI handles the heavy lifting, not the heart.

💚 AI-first ≠ AI-everything - We use it where it adds value, not just because it’s shiny.

💚 The editor’s mindset - Every output is reviewed by a real human with a red pen (and a sense of humour).

💚 Transparent & fair - We check for bias, respect privacy, and explain our choices.

💚 Constantly learning - Because AI changes faster than your morning coffee gets cold.

💚 Ethics at the core - Our tools and processes are designed to protect privacy, accuracy, and fairness.

🎥 Here’s what happened when we tried to build a brand new 3SC website, the AI-first way.

How Does AI Help Us Do More Good?

AI gives us super-speed. Humanity gives us purpose.

💚 Creative & Comms: brainstorming ideas, research, and writing first drafts (never final ones).

💚 Design & Dev: speeding up prototypes, debugging, and cleaning code.

💚 Podcast & Content: transcriptions, captions, and summaries to make stories more accessible.

💚 Strategy & Insight: finding trends, simplifying complex data, and helping teams make informed decisions.

Everything goes through human review and our trusty red pen 🖍

Our AI Pinky Promise

Person coding

Where our ethics meet their algorithm.

Because even AI needs ground rules. We take our Tech for Good mission seriously, and why we have this AI governance framework. Here’s how we keep things ethical, secure, and unapologetically human at Cube:

App-development

Privacy & Protection

Client data never touches an AI prompt. Ever. We lock it down, encrypt it, and only use pre-approved tools that pass our security and privacy checks.

People in Meeting

Human in the Loop

AI assists, it never decides. Every output is reviewed for bias, accuracy, and tone before it goes anywhere near a client or a campaign.

post-launch-support

Fairness & Inclusivity

We check for bias like it’s a bug in the code. Different perspectives make better products and fairer outcomes for everyone.

Two of the Cube team discussing a project in front of a whiteboard

Accountability

When we nail it, we celebrate. When we don’t, we own it, fix it, and share what we learned. Responsibility isn’t a department here; it’s all Cubie’s on deck for this.

two people sat in front of laptops collaborating on some testing

Reliability

We test, re-test, and validate until it works in the wild, not just in theory. If it can’t hold up under pressure, it doesn’t make the cut.

designer using a drawing board to wireframe mobile app screens by hand

Continuous Learning

AI evolves fast, and so do we. We review our policy every six months and share what we learn - wins, fails, and everything in between.

Built Different (By Design)

How we experiment, learn, and keep it 💯

The way we build has changed, the reason why hasn’t.

We’ve baked AI solutions into our design sprints, dev builds, and rubber duck sessions because it helps us move faster and think smarter without losing the human stuff that actually matters.

As always, we lead with curiosity, just with a few extra tools in the mix.Sometimes that means a prototype built in half the time. Sometimes it means AI catches something we might’ve missed. Either way, it doesn’t run the show,  it’s our trusty side-kick.

Each experiment feeds into how we build. Less guesswork, more getting it done, and even more impact 🚀

Ready to use AI for Good? So are we

(We promise we won't make you talk to a bot)

Get In Touch