Tagged: creativity
Shame on you, H.R.!
I struggle mightily with understanding why H.R. is so willingly making itself an enabler of generative A.I., specifically in the hiring of creatives, and in the displacement and replacement of creative talent.
At best, it is inconsistent and self-defeating. At worst, it is deeply immoral.
First, you put up requirements for creatives to showcase expensive art, design or writing degrees. Then, you expect them to prove their abilities by proffering extensive examples of their work experience, and providing references to back it up. Lastly, you expect to be served examples of creative work, which will presumably be judged (subjectively, but hopefully by people who are qualified to make that judgment), based on its inherent qualities. Then, at the very end, you expect them to replace all of that knowledge, all of that expertise, all of those evident qualities, with homogenized and industrialized machine output, reducing all of their varied experience to prompt writing. Why on Earth would anyone take on a mountain of student debt to learn a trade, taking pride in the quality of their work, using that pride to further the objectives of their employer, only to have their work be reduced to data entry, thereby becoming complicit in the complete erosion of their entire profession, and the depleted value of the skillset they’ve fought so long to attain?
There are profound problems inherent in equating generative A.I. with human expertise and human learning. Let’s be clear here: whatever A.I. tech advocates may say to the contrary, generative A.I. engines don’t actually learn in the true semantic sense of the word. They are not sentient, they are not able to draw conclusions and weigh them against each other, or extrapolate from what they see. There is no lateral thinking whatsoever. There is no trial-and-error, no judicious application of the many considerations going into the appropriateness and “feel” of creative work. There is no social context for the use of the output of generative A.I. These machines are trained to identify patterns, replicate them, and fuse together replicated pieces, like a high-tech meat grinder. This is a BIG difference compared to how humans work, how actual learning works, and how creativity works.
More importantly, that replication is fraught with legal, commercial and moral problems. Some examples:
TRUE LEARNING VS. REPLICATING
A human artist who learns from another artist would probably initially copy the other artist’s work, but eventually transcend those influences, and emerge with a style more their own. With generative A.I., sure, you can feed the machine examples of an artist’s work, but it doesn’t become that artist by ingesting those pieces of visual data, and it never develops aesthetic sensibilities of its own. All the A.I. learns how to do is to imitate that style, which is lightyears away from the organic inspiration and learning that happens between human artists. Moreover, the output is judged by how precisely the A.I. approximates someone else’s work, whereas a human artist is typically judged based on the uniqueness of their output. If a human artist simply replicated the work of another artist, there is a word for that: it is called counterfeiting. The counterfeiter could (and would!) be sued and be taken to court. Surely, you wouldn’t expect a company like Disney, for instance, to accept the wholesale, industrial scale duplication of artistic work they’ve spent generations refining, would you?
UNIQUE VS. COPYCAT
Then we have the field of branding and corporate identity, which would suffer immensely from the uncritical adoption of generative A.I. First of all, anything incorporated into a brand’s visual or verbal presence that has been created from replicated pieces of other creative output is obviously not unique, it goes without saying. Hence, it flies in the face of the very nature of brand building. Its value in defining a distinct identity, or catching the attention of potential customers by standing out, is inherently compromised. Second, what is not unique cannot be owned or claimed as your own: others can freely copy it, and the coherence and recognizability of a brand would suffer immensely from it. Commercially, this would devalue the brand, far more than any savings possible from using A.I.
DEVELOPING VS. vs. STEALING
There are ways of creatively developing new, improved forms of output without directly replicating something already existing, and that process is not new: it’s called R&D. The perennial truth in that field is that there are no shortcuts. If a car manufacturer wanted to develop a sportier car, they wouldn’t steal a racecar, take it apart, replicate its patented componentry, and implement it as-is. Instead, they would analyze the racecar, understand how it worked, and then apply whatever mechanical and aerodynamical principles that were observable and applicable to their product, and then iterate and test it until they found a workable set of components that produced the intended result, and were feasible to produce in an economical way. Part of that testing would involve how it felt to drive the car, a sensation that any A.I. would struggle to account for in its output. Moreover, the cheating involved in simply replicating existing components would be considered a crime; one for which the manufacturer would be taken to court and punished by the letter of the law. There’s an entire legal profession dedicated to this, and you better believe they are gearing up to incorporate the defense against abuse of A.I. into their law practices.
GROWTH VS. ATROPHY
If you, as an H.R. professional, think that you are contributing to growing the competence of your creative department by hiring people with a focus on generative A.I. usage, you have an extremely short-sighted perspective. Do you seriously think, for instance, that the creative ability to envision and visualize bespoke solutions won’t be affected if you have people do nothing but sit and feed prompts to a machine, and judge what of its output is usable…? That removing creative immersion in artistic decisionmaking will not lead to the atrophy of creative abilities in your staff…? Imagine if you had your Art Directors do nothing but conduct Google image searches all day, being fed algorithm-homogenized images for years, recycling the same visuals in a giant aesthetic echo chamber. Do you think they would ever come up with anything eye-opening or truly creative ever again? We’ve already seen the effects of algorithm-based selection of content in the increasingly isolated thought bubbles of social media: it leads to an erosion in human contact, and a depletion of human ingenuity. Instead of coming up with sentiments that more truthfully and genuinely represent people’s opinions and feelings, we are reduced to sharing and recycling memes, and using pre-defined emojis. Imagine applying that same homogenizing effect to the entire field of creative work! If that doesn’t give you pause, you seriously haven’t given it enough thought.
CONVINCING VS. CONNING
The entire field of marketing is based on the principles of persuasion: that advertising can somehow convince a potential customer to change their purchasing decision in your favor. This has, so far, been entirely dependent on human-to-human communication, as it should be. Meaning, if you are being persuaded, it was ultimately another human being who persuaded you. If they did so through illegitimate means – by exaggerating, obfuscating, lying, swindling – then that is something for which a human can be held accountable. In the case of machine marketing output, what does that accountability look like? Nobody knows, and it will take decades of legal wrangling in court to establish enough legal precedents for the law to be consider settled. In the meantime, we will see A.I.s step over the line and repeatedly lie to people, without knowing it is doing so, without its handlers knowing it did so, and without anyone being accountable. We’re already seeing that happen: very recently, a summer reading list with AI-generated content, including fake books and quotes, was published by several newspapers. This will lead to a deepening of already dangerous levels of untruth.
AGENCY AND RESPONSIBILITY
It is my firm belief that, at a certain level – say Director level and up – any professional should have a say in which tools are used to practice their trade, and how they are used. More than that, it should be part of their responsibilities. That is, in fact, part of what you are hiring them for, and more importantly, you should be hiring them to stand up for what is right, not to become unthinking tools of machine adoption. Denying professionals that agency, that choice, and that say in the execution of their own professional duties, is tantamount to a form of abuse – especially if it leads to the depletion and devaluing of their hard-earned capabilities, and the long-term dismantling of their own profession.
If you go into H.R. thinking that it’s your job to enable your employer’s abuse of their employees, I would suggest there is a deep flaw in your moral compass.
Shame on you.
The Brittleness of Ideation
I’m intensely protective of the creative process. It needs to be nurtured and cared for in its fragile infancy, before ideas are fully formed.
Ideas need to be moved forward gently and carefully, while teasing out their shape and structure. Every contributor needs to be afforded the headspace to go into all the nooks and crannies of their creativity, and explore hidden potential without hesitation or stifling judgment. This is not typically a matter of individual personality or psychological fortitude; even those who seem outwardly resilient alter their patterns of exploration based on some measure of self-doubt, external factors, or interpersonal dynamics. This usually happens imperceptively and subconsciously, and there are no guarantees that these subtle course alterations are for the better.
The creative process depends on open-mindedness and a nurturing, supportive attitude. In this regard, it is not unlike therapy – especially when the process is collective and collaborative. The realm of ideas needs to be a safe space for all participants, where they can feel secure in probing for those elusive impulses that may or may not deliver something of substance. The key response should always be “yes, and…”, or a “maybe”, as opposed to a flat “no”. The rule of thumb must be to suspend judgment and disbelief, and move forward with the assumption that every grain of sand might potentially be hiding a speck of gold. Otherwise, the risk is that viable ideas are overlooked, or modified in ways that do not fully leverage their potential. Ideas are constantly evolving, so to regard them as being in stasis is simply ignoring the inherent potential.
Issuing opinions or judgments is almost always premature while ideas are still congealing. While the instinct to judge, to limit and to reject is profoundly human – it is related to our fight-or-flight instincts – it is deeply at odds with the ideation process. Rejection is an expression of fear, whereas acceptance is an expression of love. Evaluation inserted too early in the ideation process only serves to restrict the freedom of movement of ideas before they are fully formed, especially if such evaluation is based on personal bias and hypotheticals – almost regardless of how well informed such hypotheticals may be. Only when an idea has been given a clear shape and direction is it meaningful to evaluate it, and decide how to proceed, and this “ready” state is not always immediately apparent.
Therefore, there is really no need to rush to judgment, at any stage of the ideation process. Let the process take its time. Every single decision already happens under the natural pressure of second-guessing and doubt; that is a fully normal condition in navigating the ambiguity of the unknown and unseen. We must, as creatives, strive to lessen these doubts, and not succumb to self-censorship. This is especially true in collaborative ideation, as we can unwittingly have a censoring and suppressing effect on others by merely issuing an opinion, and this often leads to the halting of forward momentum, the stifling of creativity, and/or the strangling of the flow of thought. Even those who may seem outwardly resilient to such undue influence are subject to these effects, and those who lack the necessary understanding of these psychological group dynamics are typically better removed from the process, as their presence can be quite destructive. This calls for some self-awareness: I have myself had this effect on ideation on occasion, and have had to bow out of the process for that reason.
Instead, what we need to do as creatives is to welcome, document and catalogue ideas, move on to the next idea, and then see if patterns emerge where certain concepts share similarities, or may contradict each other to the point where it eventually becomes necessary to pick a path. There is ample time for judgment at the end of this process, but to pre-empt the potential value of an idea before it has fully revealed itself is never productive. Though picking paths may give a false sense of progression, it does not enhance ideation to choose a direction too narrowly before it is well understood where a path may lead to; it only restricts options in a manner that is rarely well considered or constructive. Some paths may converge, and some may diverge, but they might all lead to meaningful destinations.
If we apply a divergent mindset, as opposed to a convergent one, we ensure that we explore more possibilities and possible combinations of ideas, and even if some of these possibilities turn out to not be applicable for the task at hand, they very often lead to new trains of thought, and spawn new ideas which may serve some other purpose at a later point. We do ourselves a disservice by rejecting them.
This is why the true value of an idea can never just be assessed within the narrow confines of one specific objective. An idea always has value, it just needs to be nurtured, appreciated and considered in the right light.
