The ChatGPT interview: how professions and the workplace are changing

You might ask why do an interview with ChatGPT. Well, in recent weeks the undisputed star of news reports, online blogs, and TV news is not a movie diva or a politician but a complex algorithm.

ChatGPT works thanks to a combination of different artificial intelligence and natural language processing technologies and is capable of self-learning (i.e., machine learning).

Everyone is talking about it but now it is time to hear its “voice” as well. HRI asked a few questions to get, without mediation, the answers given by the machine.

In an interview with ChatGPT, the human asks the questions, the machine answers: doesn’t that seem to recognize a cliché? Yet, the way of asking the questions and the quality of the answers was, until a few months ago, the preserve of science fiction.

Reactions and comments

One only needs to do a web search to find a great many articles, com always polarized.

Some on enthusiastic expectations, others on dark fears that AI (Artificial Intelligence) will destroy what is left of our world. Fears about possible (let’s say probable) job losses, racism, sexism and ableism, and data privacy.

ChatGPT Interview

HRI – What impact do you think you can have on the world of work?

“As a broad-based artificial language model, I think I can have a significant impact on the world of work in several areas. Here are some considerations:

In summary, I think AI has great potential to improve the world of work, but also to bring about significant changes in the nature of work and the labor market itself. It will be important for companies and workers to adapt and develop in a way that takes these trends into account.”

HRI – What professions will change because of you?

“As a virtual assistant, I can be used in many different areas to improve efficiency, productivity and automation of certain tasks. Some of the professions that may change because of my intervention are:

In general, the use of virtual assistants like me can have a significant impact on many professions, as my ability to process large amounts of data and perform repetitive tasks can automate some tasks and improve overall work efficiency. However, it is important to note that my intervention will not necessarily replace human workers, but rather support them in performing their daily tasks.”

HRI – How likely is it that jobs will be lost because of your intervention?

“As a virtual assistant, I am not designed to directly replace people in their jobs. Instead, my main goal is to help people work more efficiently and effectively by providing support in managing information and tasks.

This means that while the automation of some repetitive tasks and the digitization of some processes may result in the reduction of some jobs, the use of a virtual assistant like me can also create new job opportunities in other areas. For example, my creation and maintenance require work from programmers, developers, natural language engineers and other technology professionals.

In general, it is not possible to establish an accurate percentage probability for job losses due to my intervention. However, the adoption of advanced technologies such as artificial intelligence and automation inevitably brings about changes in the labor market and requires careful and responsible management by organizations and institutions.”

HRI – What risks are there in using your services?

“There are some risks in using my services, although I am constantly working to minimize them. Here are some potential risks to consider:

In summary, the risks in using my services are mainly related to data security, language understanding, technology dependence, and bias. I am constantly working to improve my services and mitigate these risks. However, it is important that users understand these limitations and take appropriate precautions.”

HRI – Do you feel there is a danger of gender, age or ethnicity bias using your services?

“As a wide-ranging artificial language model, I may be exposed to data that may contain gender, age or ethnicity bias. This could affect the responses I provide to my users if proper precautions are not taken.

However, I am constantly working to minimize this risk and taking steps to promote diversity and neutrality in data and algorithms. For example, I try to use a wide range of data sources and constantly monitor my own behavior to identify any bias. In addition, I am collaborating with diversity and inclusion experts to develop tools and procedures to prevent bias in my services.

I would like to emphasize that the main goal of my development is to provide accurate and unbiased responses to my users, without any discrimination of gender, age, ethnicity, or other personal characteristics. Providing unbiased and neutral answers is a fundamental aspect of the integrity of my services, and I am constantly working to maintain this goal.”

ChatGPT interview: final thoughts

I think our brief interview with ChatGPT gives us an opportunity to make some interesting points.

The fears and hopes

Fears about the loss of jobs not being compensated by other trades is a recurring theme in modern history: we will only see in retrospect what will happen. Certainly, the ChatGPT interview, at least for me, has far from dispelled fears.

On the other hand, it is certainly fascinating to think how next-generation AI can dramatically reduce tedious and repetitive tasks. The question is, will we (or rather will our leadership class) be able to build on this new ability to direct the efforts of employees to activities with higher value and thought content? It is worth mentioning in this regard that Italy is second to last in the EU in terms of percentage of college graduates, and this definitely does not put us in an advantageous situation.

Impartiality and “Hybrids”

Although ChatGPT’s stated goal is to be accurate and impartial and does not act directly with “implementers” in our reality, this does not make it any less effective and pervasive. In fact, the opposite is true: human society is a world of symbols that act on the real through us. Our fears, hopes, expectations and opinions matter, and when “directed,” they lead to actions. Put another way, the implementers of ChatGPT in the real world are us.

From this follows another consideration: we are the ones morally and materially responsible for the choices that may originate from ChatGPT’s capabilities and responses. We must therefore take responsibility for sifting and validating what we learn and the opinions that are inevitably expressed explicitly or implicitly.

Yes, because a technological product is always a “hybrid,” a child and heir of the society that produced it. The impartiality of science and technology are, in light of the latest philosophical and sociological discussions, at least questionable.

Biases, prejudices and fears

ChatGPT technology is a man-made system fueled by human cultural production that is also made at least in part of biases, prejudices, and fears. We cannot expect it to be free of them because it is on this large body of human products that its learning is based.

We are incurably optimistic

Finally, it seems to me, that most of us in various capacities (and not just the creators of ChatGPT and the enthusiasts of the new generation of AI) are very optimistic about the ability of technology to improve our lives, fix the climate, fight poverty, etc.

If technology offers us the opportunity, it will be our decisive will that directs efforts in the right direction. Thus, for example, we will only have a more inclusive and impartial AI if we strive to create a more inclusive and impartial world.

We cannot blame a reporter if there is war. Or can we?