Sept. 20, 2025

Journalism & News: has the internet made everyone an expert journalist? – John DeDakis

Journalism & News: has the internet made everyone an expert journalist? – John DeDakis
The player is loading ...
Journalism & News: has the internet made everyone an expert journalist? – John DeDakis

It promised unlimited reach, direct engagement, and a more informed world. But instead devalued expertise, creating a ‘clicks and outrage’ economy where speed reigns over accuracy.

To help us get the full story, we’re joined by John DeDakis, an award-winning novelist, writing coach, public speaker, and former CNN Senior Copy Editor.

https://johndedakis.com/

https://www.amazon.com/stores/author/B002BM6WM0

https://www.instagram.com/dedakisjohn/

https://www.facebook.com/john.dedakis

https://bsky.app/profile/johndedakis.bsky.social

In this episode, John DeDakis, a former CNN editor and current writing coach, discusses the profound changes in journalism brought about by the internet. He highlights the challenges of maintaining trust in news, the impact of AI and misinformation, and the importance of media literacy.

We also explore the need for public broadcasting and the role of citizen journalism in shaping the media landscape. DeDakis concludes with a hopeful outlook for the future of journalism, stressing the importance of curiosity and discernment in the pursuit of truth.

00:00 The Journey of a Journalist

02:16 The Impact of the Internet on Journalism

03:52 Skepticism and Trust in Digital News

06:25 Challenges for Editors in the Digital Age

08:48 The Shift Towards Clickbait and Investigative Journalism

11:35 The Rise of Citizen Journalism

13:55 The Role of AI in Journalism

16:37 The Future of Content Creation

19:14 The Ethics of AI and Journalism

21:19 Substack and the Future of Individual Journalism

24:08 The Evolution of Journalism in the Digital Age

30:15 The Importance of Media Literacy and Education

37:08 The Role of Public Broadcasting in Journalism

42:50 Hope for the Future of Journalism

45:18 Curiosity as a Key Skill for Writers

If you like what you hear, please follow the show, leave a review, or let us know what else to look into at ⁠⁠https://www.ruinedbytheinternet.com/⁠⁠

1
00:00:00,160 --> 00:00:01,640
Welcome to Ruined by the
Internet.

2
00:00:01,680 --> 00:00:03,960
I'm Gareth King.
Today we're asking, has the

3
00:00:03,960 --> 00:00:07,280
Internet ruined journalism?
A promised unlimited reach,

4
00:00:07,280 --> 00:00:10,640
direct engagement and a more
informed world, but instead

5
00:00:10,640 --> 00:00:13,960
devalued expertise, creating a
clicks and outrage economy where

6
00:00:13,960 --> 00:00:17,440
speed reigns over accuracy?
To help us get the full story,

7
00:00:17,440 --> 00:00:21,120
we're joined by John Dedekis, an
award winning novelist, writing

8
00:00:21,120 --> 00:00:23,800
coach and former CNN senior copy
editor.

9
00:00:28,640 --> 00:00:30,760
John, thank you so much for
joining us and welcome to the

10
00:00:30,760 --> 00:00:32,720
show.
Thanks Gareth, it's good to be

11
00:00:32,720 --> 00:00:34,520
here.
Before we get into it, can you

12
00:00:34,520 --> 00:00:37,720
tell us a bit about what you do
and the journey that led you to

13
00:00:37,720 --> 00:00:40,480
this point?
What I do now is I'm a writing

14
00:00:40,480 --> 00:00:45,120
coach, a manuscript editor, a
novelist, a public speaker, and

15
00:00:45,120 --> 00:00:49,240
I've been doing that since I
retired from CNN in 2013,

16
00:00:49,440 --> 00:00:52,680
although I was doing some of
those things while I was still

17
00:00:52,680 --> 00:00:57,760
at CNNI was a journalist for 45
years, covered the White House

18
00:00:57,760 --> 00:01:02,800
when Reagan was president, went
to CNN in 1988 in Atlanta and

19
00:01:02,800 --> 00:01:07,600
was with the network for 25
years, the last 7 as an editor

20
00:01:07,600 --> 00:01:10,080
for Wolf Blitzer on the
Situation Room.

21
00:01:10,240 --> 00:01:12,240
Yeah, wow.
So it's quite a quite an

22
00:01:12,240 --> 00:01:15,040
extensive repertoire and
background that you've got

23
00:01:15,040 --> 00:01:16,680
there.
So obviously you've been in the

24
00:01:16,680 --> 00:01:20,280
world of journalism and and
writing long enough to have seen

25
00:01:20,280 --> 00:01:23,520
the arrival of the Internet and
also its effects play out.

26
00:01:23,720 --> 00:01:27,280
From your perspective, what's
been the most fundamental change

27
00:01:27,280 --> 00:01:30,600
across the industry?
I think that the I mean the most

28
00:01:30,600 --> 00:01:34,080
fundamental change, at least for
me, because I existed before the

29
00:01:34,080 --> 00:01:38,440
Internet did the the biggest
change is the connectivity of

30
00:01:38,440 --> 00:01:41,680
it.
I mean, you and I are half a

31
00:01:41,680 --> 00:01:45,080
world away and we are able to
talk in real time.

32
00:01:45,080 --> 00:01:49,000
Being able to see each other.
I mean, that was unheard of when

33
00:01:49,000 --> 00:01:52,080
I was growing up.
And so I think that's the

34
00:01:52,080 --> 00:01:57,200
biggest thing, the reach that we
have to be able to connect with

35
00:01:57,200 --> 00:02:00,720
people around the world.
I mean, that's just spectacular,

36
00:02:00,720 --> 00:02:02,360
I think.
Yeah, absolutely.

37
00:02:02,360 --> 00:02:06,280
And I think just on that point
of reach, you know, as you said,

38
00:02:06,280 --> 00:02:08,960
it would have been unheard of,
but but looking back to that

39
00:02:08,960 --> 00:02:11,920
time, what did the initial
impact of the Internet and

40
00:02:11,920 --> 00:02:15,320
technology look like on the
industry and and what was the

41
00:02:15,320 --> 00:02:18,520
reaction at the time?
Well, in in journalism there was

42
00:02:18,520 --> 00:02:22,520
certainly a fair degree of
scepticism about the Internet.

43
00:02:22,760 --> 00:02:27,320
I mean, we used it, but we had
to be very careful about knowing

44
00:02:27,320 --> 00:02:32,080
if the information was reliable.
And so for the longest time, you

45
00:02:32,080 --> 00:02:36,080
know, the, the, the Internet was
available as sort of a tip

46
00:02:36,080 --> 00:02:38,600
service.
We would get heads up about

47
00:02:38,600 --> 00:02:42,080
things, but we would still check
it out the old way where you

48
00:02:42,280 --> 00:02:45,480
pick up the phone and call
somebody you know in authority

49
00:02:45,480 --> 00:02:49,840
at a reputable organisation, a
government agency or one of your

50
00:02:49,840 --> 00:02:54,680
contacts or something like that.
So you know, it took a while for

51
00:02:54,680 --> 00:03:00,320
the Internet to become more
insinuated into daily life.

52
00:03:00,360 --> 00:03:04,240
Yeah, interesting to hear that.
It almost generated those leads

53
00:03:04,240 --> 00:03:07,400
easier to to people which he
then still had to investigate.

54
00:03:07,400 --> 00:03:10,200
And I think one of the things
we'll probably get into as well

55
00:03:10,200 --> 00:03:12,720
as we go through this
conversation is the speed that

56
00:03:12,720 --> 00:03:15,800
everything operates at these
days and how that plays out.

57
00:03:16,000 --> 00:03:19,920
Would you say that it was the
Internet as a as a tool for

58
00:03:19,920 --> 00:03:23,360
journalism was taken seriously
right off the bat, like right

59
00:03:23,360 --> 00:03:26,320
away, or do you think the
industry was a little bit slow

60
00:03:26,320 --> 00:03:29,360
to react?
It's been a while now, but I, I

61
00:03:29,360 --> 00:03:33,440
think that there was a fair
amount of scepticism and, and I

62
00:03:33,440 --> 00:03:36,320
think you know what, I think at
reputable news organisations,

63
00:03:36,360 --> 00:03:39,680
there's still a lot of
scepticism for good reason,

64
00:03:40,040 --> 00:03:46,120
because there is just so much BS
out there and there's so much,

65
00:03:46,120 --> 00:03:49,680
there are so many lies and
conspiracy theories.

66
00:03:49,680 --> 00:03:52,120
It's just gotten like it's the
Wild West.

67
00:03:52,520 --> 00:03:55,960
And so I think, I think
reputable news organisations are

68
00:03:55,960 --> 00:04:01,480
still cautious about it and yet
it's used quite a bit as well.

69
00:04:01,560 --> 00:04:03,920
Yeah, yeah.
Especially, I mean your point

70
00:04:03,920 --> 00:04:06,440
there around reputable news
organisations.

71
00:04:06,440 --> 00:04:10,200
And I think as we know, as it
expands and and the rise of say

72
00:04:10,200 --> 00:04:13,360
citizen journalism as well, the
lines are kind of getting

73
00:04:13,360 --> 00:04:15,320
blurred.
And as I guess we've seen play

74
00:04:15,320 --> 00:04:19,760
out over years now, that trust
even in these reputable

75
00:04:20,040 --> 00:04:22,840
organisations seems to be
breaking down quite a lot too.

76
00:04:22,840 --> 00:04:26,560
So obviously that's another
challenge to to try and address

77
00:04:26,560 --> 00:04:30,120
as everything keeps going.
But I guess on that point, what

78
00:04:30,120 --> 00:04:34,480
would you say beyond maintaining
that reputation and I guess that

79
00:04:34,480 --> 00:04:37,680
trust from the audience, what
would be the biggest challenge

80
00:04:37,680 --> 00:04:42,440
for an editor trying to manage
an outlet or publication in the

81
00:04:42,440 --> 00:04:44,680
digital age?
And how would that compare to

82
00:04:44,680 --> 00:04:48,120
previously when it was kind of
just implicitly trustworthy?

83
00:04:48,360 --> 00:04:52,120
I'm not sure I can add much to
that just because journalists

84
00:04:52,160 --> 00:04:58,080
are sceptical.
And so the the problem is the

85
00:04:58,360 --> 00:05:03,280
sources that are making things
up and it's just a matter of

86
00:05:03,280 --> 00:05:06,840
trial and error.
I mean, the, the sources that

87
00:05:06,840 --> 00:05:10,760
journalists tend to rely on are
sources they trust.

88
00:05:11,560 --> 00:05:15,400
And so a lot of people are
competing for attention.

89
00:05:15,400 --> 00:05:20,960
The, the problem now is that, I
mean, you even, look, you even

90
00:05:20,960 --> 00:05:25,440
have the President of the United
States, Donald Trump, who I'll

91
00:05:25,440 --> 00:05:30,320
be honest, he, he lies
constantly, reflexively, you

92
00:05:30,320 --> 00:05:34,000
know, and, and so if he says it,
you have to be sceptical.

93
00:05:34,640 --> 00:05:38,240
But of course, if you ask him
questions, then you're, you're

94
00:05:38,240 --> 00:05:42,000
considered treasonous.
You know, it's, it's gotten to

95
00:05:42,000 --> 00:05:46,800
the point where, you know, Trump
himself has done a tremendous

96
00:05:46,800 --> 00:05:51,920
disservice to the to the
Internet because of the

97
00:05:51,920 --> 00:05:57,200
falsehoods that he spews and
then claims that the mainstream

98
00:05:57,200 --> 00:06:02,080
media is fake news.
And what he's done, it's not.

99
00:06:02,080 --> 00:06:04,120
And he's not just undermining
journalism.

100
00:06:04,360 --> 00:06:06,680
He's undermining the
intelligence community, the

101
00:06:06,680 --> 00:06:09,280
judiciary, the scientific
community.

102
00:06:09,520 --> 00:06:12,280
And that's what dictators do.
They undermine trust in

103
00:06:12,280 --> 00:06:14,320
everything so that you believe
them.

104
00:06:14,320 --> 00:06:17,640
And I mean, he said when he
first ran for president, only I

105
00:06:17,640 --> 00:06:22,120
can solve it.
Yeah, look, that's, that's an

106
00:06:22,120 --> 00:06:26,360
entire, I guess world that we
could probably spend hours going

107
00:06:26,360 --> 00:06:28,480
down.
And what it appears to me, you

108
00:06:28,480 --> 00:06:31,120
know, I'm in Australia, I'm not
on the ground in in the US

109
00:06:31,120 --> 00:06:32,960
seeing what it's like over
there.

110
00:06:33,320 --> 00:06:37,000
But what I can kind of see
around all of that is the way

111
00:06:37,000 --> 00:06:41,800
that him or his team or whoever
it is is kind of using the

112
00:06:41,800 --> 00:06:46,520
Internet in like Internet
culturally rather than an

113
00:06:46,520 --> 00:06:49,360
official sense.
And I think that not just him, I

114
00:06:49,360 --> 00:06:54,040
think all officials do it to to
varying degrees, where they've

115
00:06:54,040 --> 00:06:57,240
got a team that manages their
social media or something.

116
00:06:57,600 --> 00:07:00,240
And the team is not staffed by
people like them.

117
00:07:00,240 --> 00:07:04,240
It's staffed by young people
totally plugged into, you know,

118
00:07:04,240 --> 00:07:06,920
the way to communicate to people
online.

119
00:07:06,920 --> 00:07:11,400
So, so that communication is
formatted as Internet

120
00:07:11,400 --> 00:07:14,560
information, not kind of
official information.

121
00:07:14,560 --> 00:07:16,560
And then I think, you know,
wires get crossed somewhere

122
00:07:16,560 --> 00:07:19,120
along the line.
That's obviously one of the

123
00:07:19,800 --> 00:07:24,360
disservices that the
democratisation of information

124
00:07:24,360 --> 00:07:27,160
provides via the Internet.
One of the things that's

125
00:07:27,160 --> 00:07:30,360
happening right now in the White
House press corps is that they

126
00:07:30,360 --> 00:07:34,440
have made room for what they
call new media.

127
00:07:34,920 --> 00:07:39,760
And, and it's a rotating thing.
And they get the first question

128
00:07:39,760 --> 00:07:43,200
in the briefing.
They sit to Caroline Leavitt's

129
00:07:43,200 --> 00:07:48,800
immediate right along that that
Rose Garden wall and the New

130
00:07:48,800 --> 00:07:52,960
York Times just did a piece on
one of the, you know, the new

131
00:07:52,960 --> 00:07:55,920
media people and the guy make
stuff up.

132
00:07:55,920 --> 00:08:00,080
I mean, he he's he's he's
already got a reputation for not

133
00:08:00,080 --> 00:08:03,960
being reputable, but that's OK
for the White House because, you

134
00:08:03,960 --> 00:08:10,040
know, he says what they want him
to say and he slavishly asks

135
00:08:10,120 --> 00:08:15,560
softball questions.
It's, it's, it's so it's

136
00:08:15,560 --> 00:08:18,640
complicated because you're right
about, you know, young people,

137
00:08:18,840 --> 00:08:23,480
you know, using the technology
to get the, the message out as

138
00:08:23,480 --> 00:08:26,240
effectively as possible.
They're not necessarily dealing

139
00:08:26,240 --> 00:08:29,040
with the content.
They're dealing with the, the

140
00:08:29,040 --> 00:08:33,559
way to make sure that the reach
is as far as it needs to be.

141
00:08:33,799 --> 00:08:37,080
But you know, you've got people
in the White House, in the, in

142
00:08:37,080 --> 00:08:40,720
the in the briefing room who
aren't journalists, but they've

143
00:08:40,720 --> 00:08:44,360
got hundreds of thousands of
followers, but there's no editor

144
00:08:44,360 --> 00:08:46,560
on their shoulder going.
Where did you get that?

145
00:08:46,720 --> 00:08:49,400
How do you know that's true?
Yeah.

146
00:08:49,400 --> 00:08:52,600
I mean that's that's an
interesting point as well around

147
00:08:52,600 --> 00:08:55,680
kind of I guess the speed that
everything needs to operate.

148
00:08:55,960 --> 00:09:00,320
We know that headlines Dr
clicks, which drive revenue and

149
00:09:01,280 --> 00:09:04,000
you know, the byproduct of that,
of course, is the need to create

150
00:09:04,000 --> 00:09:06,960
that kind of viral content that
you know, is going to get a lot

151
00:09:06,960 --> 00:09:10,080
of eyes, a lot of clicks and get
that revenue to keep the the

152
00:09:10,080 --> 00:09:13,640
network or your own small
publication, whatever is going.

153
00:09:13,640 --> 00:09:17,840
But seeing them somewhat of the
model head down that path, you

154
00:09:17,840 --> 00:09:20,200
know, where everything is just,
you know, we've seen the rise of

155
00:09:20,200 --> 00:09:24,160
opinion pieces as as like, you
know, quite polarising by their

156
00:09:24,160 --> 00:09:26,760
nature for that, I guess for
that purpose.

157
00:09:27,120 --> 00:09:31,160
Does that kind of cheap form of
content, Does that mean that

158
00:09:31,280 --> 00:09:34,480
expensive investigative
journalism and writing just

159
00:09:34,480 --> 00:09:37,080
becomes too expensive to produce
for the most part?

160
00:09:37,680 --> 00:09:41,320
And and does that shift towards
clickbait and headline, You

161
00:09:41,320 --> 00:09:43,080
know, a lot of people don't read
beyond the headline.

162
00:09:43,880 --> 00:09:46,680
Is driving things through
headlines rather than content

163
00:09:46,680 --> 00:09:49,440
simply a necessity of a business
model that might be struggling?

164
00:09:49,720 --> 00:09:52,000
Yes, that's that's a real
danger.

165
00:09:52,000 --> 00:09:56,160
And what's happening is, in
fact, the Internet is probably

166
00:09:56,360 --> 00:09:59,400
responsible for a lot of
newspapers going out of business

167
00:10:00,040 --> 00:10:03,680
because they had to monetize
what they were doing.

168
00:10:03,680 --> 00:10:07,880
But people wanted free news.
And, and so they were, you know,

169
00:10:07,880 --> 00:10:11,680
a lot of a lot of newspapers
were late to put up a paywall.

170
00:10:11,680 --> 00:10:16,200
And so they're, they were kind
of caught flat footed responding

171
00:10:16,200 --> 00:10:19,120
to the Internet.
And so you're absolutely right,

172
00:10:19,200 --> 00:10:23,360
it really puts a chill in
investigative reporting because

173
00:10:23,800 --> 00:10:27,960
in order to do effective
investigative reporting, you

174
00:10:27,960 --> 00:10:31,240
have to be backed up by a news
organisation that has deep

175
00:10:31,240 --> 00:10:34,800
pockets and skilled lawyers that
can protect you.

176
00:10:35,160 --> 00:10:38,480
You know, if you are branching
out on your own and you're doing

177
00:10:38,480 --> 00:10:43,360
investigative reporting, you are
very likely going to get sued.

178
00:10:43,640 --> 00:10:46,440
And even if you are in the
right, even if truth is an

179
00:10:46,440 --> 00:10:50,400
absolute defence, someone with
deep pockets and a lot of

180
00:10:50,400 --> 00:10:54,040
patients can run out the clock
and run out your bank account.

181
00:10:54,280 --> 00:10:57,400
Even though, I mean, look at
we're we're seeing Trump shake

182
00:10:57,400 --> 00:11:01,240
down, you know, the Wall Street
Journal, CBS News, The New York

183
00:11:01,240 --> 00:11:05,080
Times, you know, he's suing, you
know, some of these people for

184
00:11:05,080 --> 00:11:08,480
$20 billion and they're
settling.

185
00:11:08,680 --> 00:11:12,000
And these are big news
organisations, you know, they're

186
00:11:12,000 --> 00:11:15,160
not settling for 20 billion,
they're settling for 16,000,000.

187
00:11:15,400 --> 00:11:17,840
Just to kind of get it off their
plate now, of course.

188
00:11:17,840 --> 00:11:21,120
And look, it's, it's, I guess
just for context, it's very

189
00:11:21,120 --> 00:11:26,200
different over here in terms of
the litigious nature of, of

190
00:11:26,200 --> 00:11:27,960
things.
I think over here our legal

191
00:11:27,960 --> 00:11:32,360
system doesn't allow, well, not
from what I can see anyway, like

192
00:11:32,360 --> 00:11:35,920
that level of, of suing over of
kind of everything.

193
00:11:36,360 --> 00:11:40,200
But you said you said something
there around orgs, you know, and

194
00:11:40,200 --> 00:11:43,280
publications being slow to put
up a paywall around their

195
00:11:43,280 --> 00:11:46,000
content.
Why did they originally just

196
00:11:46,000 --> 00:11:48,600
decide to start giving away the
content for free?

197
00:11:48,640 --> 00:11:51,640
What was the strategy behind
that and how did they imagine

198
00:11:51,640 --> 00:11:53,720
they could capitalise on that
economically?

199
00:11:54,000 --> 00:11:57,320
It's hard for me to, to know for
sure, but my hunch is they still

200
00:11:57,320 --> 00:12:01,960
had the, the physical newspaper
and they, I don't think realised

201
00:12:01,960 --> 00:12:05,520
that, you know, people were
moving to the Internet.

202
00:12:05,520 --> 00:12:09,360
And I think that they still
trusted in the history that

203
00:12:09,360 --> 00:12:13,120
people will still gravitate
towards the actual paper.

204
00:12:13,120 --> 00:12:17,000
And I think they just were slow
to realise the tectonic shift

205
00:12:17,240 --> 00:12:19,560
that took place in news
consumption.

206
00:12:19,800 --> 00:12:21,240
And that, that makes total
sense.

207
00:12:21,240 --> 00:12:24,720
And you know, on that point, how
the how do you think the rise of

208
00:12:24,720 --> 00:12:27,720
that citizen journalism that we
we mentioned a couple of minutes

209
00:12:27,720 --> 00:12:30,840
ago, which is obviously fueled
by the Internet and the ease it

210
00:12:30,840 --> 00:12:33,760
is to publish and and share your
own stuff.

211
00:12:34,280 --> 00:12:36,760
How do you think that changed
the way traditional media

212
00:12:36,760 --> 00:12:40,040
outlets do their jobs?
Like now, with the ability for

213
00:12:40,040 --> 00:12:45,400
anyone to curate their own
unique interest, News feed is

214
00:12:45,400 --> 00:12:49,920
the one size fits all, let's say
broadcast model a thing of the

215
00:12:49,920 --> 00:12:52,560
past and like how much future do
you reckon it has?

216
00:12:52,960 --> 00:12:56,680
That's an intriguing question.
I think that the search for

217
00:12:56,680 --> 00:13:00,960
truth, I don't think that the
principles involved with that

218
00:13:00,960 --> 00:13:04,600
have gone away or will go away.
You know, you still need to

219
00:13:04,640 --> 00:13:07,320
verify your sources.
If you're getting something

220
00:13:07,320 --> 00:13:11,240
anonymously, you need to be able
to confirm it with two other

221
00:13:11,240 --> 00:13:14,320
sources at least.
There are still reputable news

222
00:13:14,320 --> 00:13:17,720
organisations, The Associated
Press, Reuters, The New York

223
00:13:17,720 --> 00:13:21,680
Times, the Guardian, BBC, just
to name a few.

224
00:13:21,880 --> 00:13:27,760
So I, I still think that the
fundamentals of journalism still

225
00:13:27,840 --> 00:13:31,520
still exist.
But the problem, of course, is

226
00:13:31,520 --> 00:13:33,640
that even the Russian
government, I mean, the

227
00:13:33,640 --> 00:13:38,120
military, they are, they have,
they have, what are they troll

228
00:13:38,120 --> 00:13:39,560
farms?
You know, they, they are

229
00:13:39,560 --> 00:13:43,800
planting false information.
You know, that that is something

230
00:13:43,800 --> 00:13:47,720
they're doing on an industrial
strength level, on a state

231
00:13:47,720 --> 00:13:50,920
level.
And, and it's, it's weaponizing

232
00:13:50,920 --> 00:13:53,840
falsehoods and that's hard to
fight against.

233
00:13:53,840 --> 00:13:56,120
It's hard to fight against.
No, absolutely.

234
00:13:56,120 --> 00:13:59,000
And I think that, you know,
that's something that I'm sure

235
00:13:59,000 --> 00:14:01,320
you you'd understand that comes
up here as well.

236
00:14:01,320 --> 00:14:05,240
You know, anytime there's an
election or anything to do with

237
00:14:05,240 --> 00:14:09,520
government going on, it's this
kind of it's whether whether

238
00:14:09,520 --> 00:14:12,520
it's it's real or not.
It's almost the go to now that

239
00:14:12,520 --> 00:14:16,200
they're they're these Russian
troll farms are doing these kind

240
00:14:16,200 --> 00:14:18,560
of interference jobs.
And I'm sure that they're not

241
00:14:18,560 --> 00:14:20,760
the only ones.
Like I'm sure basically every

242
00:14:20,760 --> 00:14:24,440
nation is probably.
China as well, you know, and one

243
00:14:24,440 --> 00:14:27,480
thing you haven't mentioned is
the is artificial intelligence,

244
00:14:27,480 --> 00:14:28,600
I mean.
Yes, yes, the.

245
00:14:28,640 --> 00:14:32,120
Deep fakes that are becoming
much more sophisticated.

246
00:14:32,280 --> 00:14:34,160
We can segue into that now
because that's a.

247
00:14:34,160 --> 00:14:36,120
It's a great, great thing to
talk about.

248
00:14:36,120 --> 00:14:39,720
I remember, you know, when that
must be a few years ago now,

249
00:14:39,720 --> 00:14:43,680
playing with the first kind of
image generators and they were

250
00:14:43,960 --> 00:14:46,280
terrible.
You know, they were laughable at

251
00:14:46,280 --> 00:14:47,600
what you would get back from
them.

252
00:14:48,160 --> 00:14:53,240
And then seeing the first
examples of, of when you could

253
00:14:53,240 --> 00:14:57,760
kind of make an image of
somebody doing something and

254
00:14:57,760 --> 00:15:01,120
then when the first video came
out where you could just take an

255
00:15:01,120 --> 00:15:03,480
existing video of put someone
else's face on it.

256
00:15:04,120 --> 00:15:06,920
Even then, you know, I remember
having conversations with people

257
00:15:06,920 --> 00:15:09,600
around this is going to be
really bad.

258
00:15:09,680 --> 00:15:14,240
Like once this kind of starts
exponentially increasing quality

259
00:15:14,240 --> 00:15:16,280
wise.
You know, it's so funny.

260
00:15:16,280 --> 00:15:19,000
We've, we've spent the last
decade.

261
00:15:19,000 --> 00:15:22,640
You, you mentioned the, the term
fake news just before and

262
00:15:22,960 --> 00:15:26,080
whether anyone likes it or not,
that has just become part of the

263
00:15:26,080 --> 00:15:28,040
vernacular for, for everybody
now.

264
00:15:28,320 --> 00:15:30,880
And, and, and it's kind of just
refers to everything that

265
00:15:30,880 --> 00:15:33,640
somebody doesn't like.
They can kind of call it that.

266
00:15:33,920 --> 00:15:36,600
And, and it's so it's so
interesting to me that we've

267
00:15:36,600 --> 00:15:42,680
spent so much of the last decade
fighting against fake news and,

268
00:15:42,680 --> 00:15:45,160
and things that are
misinformation, disinformation,

269
00:15:45,160 --> 00:15:50,320
etcetera, to now be barrelling
at super speed into a digital

270
00:15:50,320 --> 00:15:55,920
world full of stuff that's fake
and we'll have no way of

271
00:15:55,920 --> 00:15:58,560
telling.
And it's just quite an

272
00:15:58,600 --> 00:16:02,680
interesting irony that all of
that fight against fake stuff

273
00:16:03,160 --> 00:16:06,680
seems to have fallen by the
wayside as we now embrace, you

274
00:16:06,680 --> 00:16:09,240
know, generative AI and other
forms of AI.

275
00:16:10,120 --> 00:16:12,280
How do you think that's going to
happen and play out?

276
00:16:12,520 --> 00:16:15,920
Well, here's the thing.
I mean, technology is morally

277
00:16:15,920 --> 00:16:21,360
neutral and the, the, the
problem is the people and, and

278
00:16:21,360 --> 00:16:25,400
what people do with it.
And, you know, going forward, I

279
00:16:25,400 --> 00:16:28,760
mean, I think that one of the
points you made about AI being

280
00:16:28,760 --> 00:16:32,240
so sophisticated that we won't
be able to tell the real from

281
00:16:32,240 --> 00:16:35,080
the fake.
But I think that there are ways

282
00:16:35,080 --> 00:16:38,640
that there might be able to put
watermarks on things or, you

283
00:16:38,640 --> 00:16:41,720
know, a digital watermark for
authenticity's sake.

284
00:16:41,960 --> 00:16:45,240
I mean, this is way beyond my
understanding or ability to deal

285
00:16:45,240 --> 00:16:47,480
with.
And, you know, in, in this

286
00:16:47,480 --> 00:16:50,400
country, there's, you know, the
1st Amendment freedom of speech.

287
00:16:50,760 --> 00:16:54,000
And so, you know, you, you
really want to be very careful

288
00:16:54,000 --> 00:16:57,800
about fiddling with content
because that was that was always

289
00:16:57,800 --> 00:16:59,920
the case.
Even when we're talking about

290
00:16:59,920 --> 00:17:03,960
the printing press, you could
still lie, you know, and use a

291
00:17:03,960 --> 00:17:06,560
printing press to do it and
reach a lot of people.

292
00:17:06,560 --> 00:17:09,160
So, you know, lying is nothing
new.

293
00:17:09,560 --> 00:17:13,520
It's just that the technology in
the pipeline is more

294
00:17:13,520 --> 00:17:17,800
sophisticated, but the
responsibility is still on us to

295
00:17:17,800 --> 00:17:20,280
be discerning.
There's no surprise that there's

296
00:17:20,280 --> 00:17:24,599
lying out there, so that means
we've just got to be alert to it

297
00:17:24,599 --> 00:17:26,440
and not just swallow everything
whole.

298
00:17:26,960 --> 00:17:29,680
Yeah, no, totally.
And I think that your your point

299
00:17:29,680 --> 00:17:32,320
you raised there about
watermarks or or something like

300
00:17:32,320 --> 00:17:34,280
that.
You know, I have seen certain

301
00:17:34,280 --> 00:17:37,240
pieces of, of content that might
be shared on various platforms

302
00:17:37,240 --> 00:17:40,240
that will have kind of a
watermark identifier built in.

303
00:17:40,240 --> 00:17:43,520
People do want that, you know,
because we've seen the amount of

304
00:17:43,560 --> 00:17:48,160
what they call slop out there in
terms of text and and and

305
00:17:48,160 --> 00:17:52,000
writing is insane.
And everybody, you know, they

306
00:17:52,000 --> 00:17:55,320
have these debates around how
you can tell formats.

307
00:17:55,320 --> 00:17:57,440
It's got a certain tone,
etcetera.

308
00:17:57,440 --> 00:18:00,360
Like the use of an M dash.
She's like a big one, but that's

309
00:18:00,360 --> 00:18:02,760
short form content.
Obviously you're you've got a

310
00:18:02,760 --> 00:18:06,480
background in journalism, but
you're now a longer form writer

311
00:18:06,480 --> 00:18:08,600
in the world of novels.
And I make it up.

312
00:18:09,120 --> 00:18:11,120
Yeah, yeah.
But I guess that's coming from

313
00:18:11,120 --> 00:18:14,840
your from your, your own mind.
How do these are these tools,

314
00:18:15,520 --> 00:18:20,160
are they starting to impact that
world at all or are they OK?

315
00:18:20,160 --> 00:18:23,120
In what ways?
I want to say, I want to say one

316
00:18:23,120 --> 00:18:25,640
thing, though, to go back to,
you know, the future.

317
00:18:25,840 --> 00:18:29,600
I think a lot can be said for
the education system.

318
00:18:29,960 --> 00:18:33,560
I think that there really needs
to be not just, you know, in

319
00:18:33,560 --> 00:18:37,680
education where people learn how
to use it, but I think that

320
00:18:37,680 --> 00:18:41,280
there needs to be some classes
in, you know, being more

321
00:18:41,280 --> 00:18:45,480
sophisticated in using it
responsibly and being able to

322
00:18:45,480 --> 00:18:47,320
discern what's real and what's
fake.

323
00:18:47,400 --> 00:18:51,200
So I think education in
elementary school on up, you

324
00:18:51,200 --> 00:18:54,920
know, I think is, is important,
but as far as AI is concerned,

325
00:18:54,960 --> 00:18:58,280
that's something I've been
playing with because I teach a

326
00:18:58,280 --> 00:19:00,120
lot.
I do a lot of writing classes.

327
00:19:00,560 --> 00:19:03,160
And so the, the question comes
up a lot.

328
00:19:03,160 --> 00:19:07,920
And so I've, you know, I've
fiddled with AIA little bit and

329
00:19:09,120 --> 00:19:11,640
it's here to stay.
And I found that it can be

330
00:19:11,640 --> 00:19:14,920
useful as a research tool as
long as you get, you know,

331
00:19:15,000 --> 00:19:19,600
sources that you can check out.
But you know, AI isn't really

332
00:19:19,600 --> 00:19:21,920
creative.
It just regurgitates

333
00:19:21,960 --> 00:19:24,160
information.
And I actually tried that.

334
00:19:24,440 --> 00:19:27,640
I had to write a short story,
which is not my specialty.

335
00:19:27,880 --> 00:19:29,880
And so I plugged in the
parameters.

336
00:19:29,880 --> 00:19:32,920
I need a short story, 5000
words, this style.

337
00:19:32,920 --> 00:19:35,040
This is the title.
These are the elements.

338
00:19:35,440 --> 00:19:39,760
And it's spit out something in
nanoseconds and it sucked it.

339
00:19:39,760 --> 00:19:43,760
Suck.
And that then was a reminder

340
00:19:43,760 --> 00:19:47,160
that I hadn't lost my chops and
that, you know, there is

341
00:19:47,160 --> 00:19:50,520
something to say for the human
element here.

342
00:19:50,800 --> 00:19:54,240
And it is definitely, I mean,
there's a lawsuit in the US

343
00:19:54,240 --> 00:19:57,320
about copyright infringement
because, you know, people's

344
00:19:57,320 --> 00:20:00,480
books are being used to train
AI.

345
00:20:00,680 --> 00:20:04,080
But, you know, ChatGPT is not
buying these people's novels.

346
00:20:04,080 --> 00:20:07,480
They're just stealing them.
And and that and that's still

347
00:20:07,480 --> 00:20:10,640
going through the courts.
So copyright infringement is an

348
00:20:10,640 --> 00:20:12,480
issue.
But, you know, it's almost a

349
00:20:12,480 --> 00:20:15,680
moot point because the damage
has already been done, Although

350
00:20:15,680 --> 00:20:18,800
it could protect, it could
protect people who are now

351
00:20:18,800 --> 00:20:21,760
writing, but their stuff hasn't,
you know, been published yet.

352
00:20:22,520 --> 00:20:24,000
Yeah.
And on the, on the kind of, I

353
00:20:24,000 --> 00:20:27,440
guess copyright infringement
stuff, as you've alluded to

354
00:20:27,440 --> 00:20:31,080
there, it is just pulling from
all information available to it

355
00:20:31,080 --> 00:20:34,680
and kind of summarising it,
which if you, if you play that

356
00:20:34,680 --> 00:20:37,520
out over a long enough time, the
more it does that, the more of

357
00:20:37,520 --> 00:20:40,360
itself it's putting out there
and then the more of itself it's

358
00:20:40,360 --> 00:20:43,160
kind of drawing from.
So it should theoretically just

359
00:20:43,160 --> 00:20:47,320
come to this ultimately beige
point that it's just nothing.

360
00:20:47,440 --> 00:20:49,600
You know what I mean?
Once it loses, as you said, all

361
00:20:49,600 --> 00:20:53,880
that humanness, that's, yeah,
not very exciting to think

362
00:20:53,880 --> 00:20:56,160
about.
I would bet though that AI is

363
00:20:56,160 --> 00:20:59,880
becoming is going to become much
more sophisticated to the point

364
00:20:59,880 --> 00:21:03,920
where it really can replicate
the human thought and the

365
00:21:03,920 --> 00:21:06,360
emotions.
I read or I heard a story

366
00:21:06,360 --> 00:21:12,720
recently that you know, unstable
people are using ChatGPT or or

367
00:21:12,960 --> 00:21:16,960
you know, generative AI to to be
their therapist and that it

368
00:21:16,960 --> 00:21:19,480
actually can talk them into
committing suicide.

369
00:21:19,600 --> 00:21:22,280
Yeah, I've seen a couple of the
same stories and that's very,

370
00:21:22,280 --> 00:21:28,480
very scary that not only are we
so insularised, I don't know if

371
00:21:28,480 --> 00:21:32,000
that's a word, but we're so
interested now, you know, our

372
00:21:32,000 --> 00:21:35,040
devices and things and our
digital lives where we're not

373
00:21:35,040 --> 00:21:38,600
interacting with real people
offline, that turning to a

374
00:21:39,160 --> 00:21:45,760
screen for therapy is it seems
like a very dystopian next step.

375
00:21:45,840 --> 00:21:48,960
And then seeing as you said,
these stories of these these

376
00:21:48,960 --> 00:21:54,120
young people ending up kind of
committing suicide be because of

377
00:21:54,120 --> 00:21:56,920
what it said or whether it's
because of what it said directly

378
00:21:56,920 --> 00:21:59,200
or not.
It's just a so tragic.

379
00:21:59,200 --> 00:22:02,920
And hopefully there gets to be
some guardrails and, you know,

380
00:22:02,920 --> 00:22:05,760
regulations around that because
that's awful.

381
00:22:05,760 --> 00:22:09,080
If you were a nefarious actor,
you could easily sabotage that

382
00:22:09,080 --> 00:22:13,640
and, and commit some kind of OP.
But that's, that's a very dark

383
00:22:13,640 --> 00:22:17,040
place to go down.
But what I did want to talk to

384
00:22:17,040 --> 00:22:20,880
you around, as you said, those
large language models and, and

385
00:22:20,880 --> 00:22:23,480
generative AI to, to make
pieces.

386
00:22:23,480 --> 00:22:28,240
If we rewind back to the context
of, say a, let's say, a book

387
00:22:28,240 --> 00:22:32,600
publisher or even a newsroom or,
or anything that deals in words,

388
00:22:34,280 --> 00:22:36,880
theoretically, could they get to
a point where you could be

389
00:22:36,880 --> 00:22:42,000
running just agentic AI systems
producing all the content

390
00:22:42,000 --> 00:22:44,080
drawing from what's happening
out there in the world?

391
00:22:44,400 --> 00:22:48,680
And so you could be running an
entirely fake newsroom or, or

392
00:22:48,680 --> 00:22:52,800
book publisher.
Yeah, awesome.

393
00:22:54,200 --> 00:22:58,280
Yeah, have a nice day.
Yeah, God, that sounds awful.

394
00:22:58,760 --> 00:23:02,120
OK, another, another thing, if
we go back to that citizen

395
00:23:02,120 --> 00:23:04,560
journalism, I'm sure you've
heard of platforms like

396
00:23:04,560 --> 00:23:07,960
Substack, right?
Would you say that on, on an

397
00:23:07,960 --> 00:23:11,120
individual level that the
Substack model is potentially A

398
00:23:11,120 --> 00:23:15,600
viable future for journalism?
Or is it just a niche where I

399
00:23:15,600 --> 00:23:18,360
guess if you've got a big enough
name and platform already, you

400
00:23:18,360 --> 00:23:21,280
can capitalise on that?
Oh, well, I think you you can

401
00:23:21,280 --> 00:23:23,040
definitely do that.
I mean, there are a couple of

402
00:23:23,040 --> 00:23:26,240
people who left the Washington
Post and, you know, started

403
00:23:26,240 --> 00:23:29,040
their own Substack.
I've always seen Substack is

404
00:23:29,040 --> 00:23:34,600
more of a newsletter kind of
thing, but you know, technology

405
00:23:34,600 --> 00:23:36,440
evolves.
I don't know if I don't see it

406
00:23:36,440 --> 00:23:40,680
necessarily replacing, you know,
day to day journalism, but it

407
00:23:40,680 --> 00:23:44,520
certainly supplements it and it
can be a useful, valuable

408
00:23:44,520 --> 00:23:49,080
contribution to the search for
truth that and I mean, that's

409
00:23:49,080 --> 00:23:51,400
the thing.
It's, I think that anything is

410
00:23:51,400 --> 00:23:56,200
possible and that it can be used
for good or for I'll.

411
00:23:56,720 --> 00:24:00,880
And it goes back to being
discerning and being able to use

412
00:24:00,880 --> 00:24:03,480
things responsibly with the
understanding that there are

413
00:24:03,480 --> 00:24:06,760
going to be a lot of people who
don't use it responsibly or use

414
00:24:06,760 --> 00:24:10,160
it nefariously or even in an
evil with evil intent.

415
00:24:10,800 --> 00:24:15,400
But that's been the human
condition since the beginning.

416
00:24:15,680 --> 00:24:18,280
It's it's just more
sophisticated now.

417
00:24:18,560 --> 00:24:21,080
Yeah, on that point, you said
something.

418
00:24:21,080 --> 00:24:23,880
They obviously the search for
truth and you got to be a bit

419
00:24:23,880 --> 00:24:27,600
discerning around whether it's
your sources or what you're

420
00:24:27,600 --> 00:24:30,280
producing.
A little while ago we touched on

421
00:24:30,280 --> 00:24:34,320
kind of how the Internet
affected, say, newsrooms and

422
00:24:34,320 --> 00:24:38,200
journalism right at the get go.
How has that changed over time

423
00:24:38,200 --> 00:24:42,680
now that there's this 24/7 news
cycle and just constant demand

424
00:24:42,680 --> 00:24:45,800
for new stuff and the speed that
everything operates at?

425
00:24:45,960 --> 00:24:50,360
Well, in in some ways, it's it's
very useful because one of the

426
00:24:50,360 --> 00:24:53,920
things I noticed when I was at
CNN is that, you know, we could,

427
00:24:53,920 --> 00:24:57,520
it was hard for us to do
interviews with people who

428
00:24:57,520 --> 00:25:01,200
didn't live in a Bureau city.
In other words, you had to get a

429
00:25:01,200 --> 00:25:04,640
camera in front of them.
And, you know, that took time,

430
00:25:04,640 --> 00:25:08,120
it took effort, travel, all that
kind of stuff.

431
00:25:08,120 --> 00:25:11,960
So, you know, the newsmakers
were in the big cities near

432
00:25:11,960 --> 00:25:15,520
where we had a camera.
Well, now with the Internet, we

433
00:25:15,520 --> 00:25:18,400
can hook U to somebody in
Function Junction, Utah.

434
00:25:18,560 --> 00:25:24,040
And that's great because the
technology allows us to reach

435
00:25:24,040 --> 00:25:28,080
more people and draw from the
expertise and the experience of

436
00:25:28,080 --> 00:25:32,640
people who aren't just in the
elite gatekeeping kinds of

437
00:25:32,640 --> 00:25:35,600
places.
So I think that's a value you

438
00:25:35,600 --> 00:25:37,720
have now.
You know, everybody who's got a

439
00:25:37,720 --> 00:25:41,800
cell phone, cell phones are very
sophisticated in that they have

440
00:25:41,800 --> 00:25:45,360
that camera function and the
video function.

441
00:25:45,600 --> 00:25:50,200
So people can go live during a
traffic stop that goes South

442
00:25:50,200 --> 00:25:53,440
and, you know, people see it in
real time badly.

443
00:25:53,440 --> 00:25:57,920
However, mass shooters have also
live streamed their crimes.

444
00:25:57,920 --> 00:26:01,720
It's sick.
But just because the ability

445
00:26:02,160 --> 00:26:05,360
that is there doesn't make the
Internet sick.

446
00:26:05,680 --> 00:26:10,040
It means that people misuse it.
But as far as journalism is

447
00:26:10,040 --> 00:26:14,840
concerned, it's been wonderful
in terms of getting video fast

448
00:26:15,080 --> 00:26:18,200
from people right on the scene.
You know, if you've got, if, you

449
00:26:18,200 --> 00:26:22,720
know, when Trump was shot in
Butler, PA, everybody had it on

450
00:26:22,720 --> 00:26:25,200
tape.
You know, I mean, you, you get

451
00:26:25,200 --> 00:26:27,720
so many different angles of the
same thing.

452
00:26:28,240 --> 00:26:31,520
And that is helpful to law
enforcement as well.

453
00:26:31,760 --> 00:26:35,360
Yeah, Look, I think that that as
well, as you said, if

454
00:26:35,360 --> 00:26:38,360
theoretically any event that
happens, you've got somebody on

455
00:26:38,360 --> 00:26:41,360
the scene.
And and we know that people not

456
00:26:41,360 --> 00:26:44,440
even interested in producing
news, they're just interested in

457
00:26:44,440 --> 00:26:47,320
getting likes online, putting a
video out there.

458
00:26:47,320 --> 00:26:51,960
And, and one of the things that
we've all noticed is whether big

459
00:26:51,960 --> 00:26:56,480
or small doesn't really matter.
Sources of news do scour the

460
00:26:56,480 --> 00:27:00,280
Internet and and social media
platforms to find those stories

461
00:27:00,920 --> 00:27:03,920
that might originate on
someone's Instagram or TikTok

462
00:27:03,920 --> 00:27:06,480
account or as you said, a video
just shared somewhere.

463
00:27:06,600 --> 00:27:09,840
How how much of that goes on?
Would you say there's people

464
00:27:09,840 --> 00:27:13,880
that that's their job just to
like be trolling the Internet

465
00:27:13,880 --> 00:27:16,840
for things like that that can
turn into stories or or do you

466
00:27:16,880 --> 00:27:20,720
hear the story and then find?
I don't think, I don't think

467
00:27:20,720 --> 00:27:22,680
news organisations are doing
that.

468
00:27:22,680 --> 00:27:25,280
Again, we're talking reputable
news organisations.

469
00:27:25,520 --> 00:27:27,600
I don't think that's where they
get their news.

470
00:27:27,840 --> 00:27:31,320
Now, there's a friend of mine in
Baltimore when the, you know,

471
00:27:31,320 --> 00:27:34,120
the Key Bridge was hit by a
barge and it collapsed.

472
00:27:34,120 --> 00:27:38,440
There was video, there was there
was surveillance video and they

473
00:27:38,440 --> 00:27:41,480
had to go through a rigorous
check to make sure they had to

474
00:27:41,480 --> 00:27:44,600
check with, you know, the
police, you know, is this is

475
00:27:44,600 --> 00:27:47,600
what we see is this did this
happen?

476
00:27:48,280 --> 00:27:52,680
And so that then means though,
that there are people out in the

477
00:27:52,680 --> 00:27:56,560
hinterlands who will, you know,
send you a video and say, isn't

478
00:27:56,560 --> 00:28:00,360
this amazing?
And it is amazing, but it's, is

479
00:28:00,360 --> 00:28:02,640
it true?
And so I think again, the

480
00:28:02,640 --> 00:28:06,000
reputable news organisations,
they're not searching for that

481
00:28:06,000 --> 00:28:08,320
kind of stuff because the stuff
comes at them.

482
00:28:09,040 --> 00:28:11,960
One of the things I'd like to
see, and maybe they actually

483
00:28:11,960 --> 00:28:17,960
exist, is look, if you've got a
video of some event that you

484
00:28:17,960 --> 00:28:21,520
know is stunning, but there's
no, you know, no one else

485
00:28:21,520 --> 00:28:25,120
happened to be there with a
camera, did it really happen or

486
00:28:25,120 --> 00:28:27,560
is it fake?
I would think there need to be

487
00:28:27,560 --> 00:28:33,720
people in newsrooms that are
able to tell if a news event is

488
00:28:33,720 --> 00:28:35,840
really a news event.
There have been a couple of

489
00:28:35,840 --> 00:28:39,080
cases where Fox News, which is
sort of a very conservative

490
00:28:39,080 --> 00:28:43,000
network in the US, sort of
Trumps personal, almost personal

491
00:28:43,000 --> 00:28:45,960
website.
They were giving stories about,

492
00:28:46,000 --> 00:28:50,240
you know, burning down cities,
but the video was from something

493
00:28:50,240 --> 00:28:53,480
entirely different.
But, you know, you got to be

494
00:28:53,480 --> 00:28:57,920
able to show where the original
came from, show how they're

495
00:28:57,920 --> 00:28:59,680
using it.
And I mean, that take.

496
00:28:59,680 --> 00:29:01,280
That's Investigative Journal.
Yeah.

497
00:29:01,280 --> 00:29:03,240
Look, that's, that's a great
point.

498
00:29:03,240 --> 00:29:07,160
And I think that that is one of
the problems with this.

499
00:29:07,160 --> 00:29:09,520
And and you know, as we, as we
mentioned a little while ago,

500
00:29:09,520 --> 00:29:13,520
like a lot of people are not
doing any deeper digging on

501
00:29:13,520 --> 00:29:16,080
anything.
You know, they see the video,

502
00:29:16,120 --> 00:29:18,560
that's the truth.
They see a claim, that's the

503
00:29:18,560 --> 00:29:20,280
truth.
You know, I've seen things like

504
00:29:20,280 --> 00:29:22,960
that, even those US ones that
you're talking about, like I

505
00:29:22,960 --> 00:29:25,680
might have seen it in my feed.
And I'm like, this just seems

506
00:29:25,680 --> 00:29:28,680
absurd, whether it's about fires
or something else.

507
00:29:29,200 --> 00:29:34,280
And within a minute of further,
you know, amateur investigation

508
00:29:34,280 --> 00:29:36,240
that I can do, I can find that
it's fake.

509
00:29:36,240 --> 00:29:40,000
But let's be honest, the vast
majority of people aren't doing

510
00:29:40,000 --> 00:29:41,120
that.
That's true.

511
00:29:41,120 --> 00:29:46,040
It's totally understandable how
this stuff spreads and I guess.

512
00:29:46,120 --> 00:29:49,720
Which which goes back to, which
goes back to my feeling that the

513
00:29:49,720 --> 00:29:52,760
education system needs to step
up as well.

514
00:29:52,960 --> 00:29:56,760
Because people are, you know,
they're sheep in many ways.

515
00:29:56,760 --> 00:29:58,600
They'll believe whatever they
want to believe.

516
00:29:59,000 --> 00:30:03,840
And that when you have an
informed electorate, when you

517
00:30:03,840 --> 00:30:07,120
know, civics is taught, when
science is taught, you know, the

518
00:30:07,120 --> 00:30:11,080
education system I think really
needs to step up to let people

519
00:30:11,080 --> 00:30:16,880
know that at a, at a young age,
that, you know, the dangers of

520
00:30:16,880 --> 00:30:20,040
falsehoods, the consequences of
that.

521
00:30:20,320 --> 00:30:23,480
I mean, I, I think we can turn
it around if we educate people.

522
00:30:23,480 --> 00:30:26,440
I mean, that's not a, that's not
meant to be a panacea, but I

523
00:30:26,440 --> 00:30:31,240
think it is at least one place
where we can start to make a

524
00:30:31,240 --> 00:30:33,440
difference.
Yeah, No, absolutely.

525
00:30:33,440 --> 00:30:36,320
I totally agree with you there.
That kind of teaching digital,

526
00:30:36,320 --> 00:30:40,480
digital literacy, I mean, look,
if you're in, if you're a 14

527
00:30:40,480 --> 00:30:43,680
year old kid now, you, you don't
know anything other than having

528
00:30:43,680 --> 00:30:46,640
a the supercomputer in your hand
at all times.

529
00:30:46,640 --> 00:30:48,680
So you know how to use all of
this stuff.

530
00:30:48,680 --> 00:30:53,080
But it's like, as you said,
being able to discern the

531
00:30:53,080 --> 00:30:56,560
objective truth versus the
subjective truth is, is a skill

532
00:30:56,560 --> 00:31:00,800
that I know it'd be interesting
to see whether people want that

533
00:31:00,800 --> 00:31:04,240
or they want their own opinions,
whatever they are to to be

534
00:31:04,240 --> 00:31:06,920
nurtured and kind of coddled in
that regard.

535
00:31:07,040 --> 00:31:09,240
You know, you mentioned Fox News
there and, and you've got

536
00:31:09,240 --> 00:31:11,920
publications on the other end.
And look, everybody's got their

537
00:31:12,080 --> 00:31:14,720
bias and, and kind of angle to
different things.

538
00:31:14,720 --> 00:31:19,440
But would you say that
journalism is in better shape

539
00:31:19,440 --> 00:31:24,240
when when you've got a strong
public broadcaster that needs to

540
00:31:24,480 --> 00:31:27,760
appeal to a broad audience
representing the entire nation?

541
00:31:27,760 --> 00:31:31,000
Like in Australia we've got the
ABC and so obviously being

542
00:31:31,000 --> 00:31:34,680
publicly funded, they need to
appear right down the middle.

543
00:31:34,720 --> 00:31:37,680
Now, that said, I don't think
there's a single person in

544
00:31:37,680 --> 00:31:40,200
Australia that thinks the ABC is
that.

545
00:31:40,200 --> 00:31:43,560
Everybody thinks that it's
catering to the views that they

546
00:31:43,560 --> 00:31:46,160
don't like, which is probably an
indicator that they're actually

547
00:31:46,160 --> 00:31:50,080
doing quite a good balanced job.
But those publicly funded

548
00:31:50,080 --> 00:31:53,200
broadcasters that do need to
play right down the middle, how

549
00:31:53,200 --> 00:31:57,240
do you think that they keep the
truth in journalism alive?

550
00:31:57,760 --> 00:32:00,280
Well, the problem is that Trump
is defunding them.

551
00:32:01,280 --> 00:32:03,800
And so that, you know, that's
one of the problems we have.

552
00:32:03,800 --> 00:32:07,440
It's not that, you know, they
never were, well, I shouldn't

553
00:32:07,440 --> 00:32:11,240
say never, but you know, at this
point it's not publicly funded.

554
00:32:11,240 --> 00:32:15,280
There's, there's a public stream
of money, but it's, it's a small

555
00:32:15,280 --> 00:32:18,760
percentage, but it's, but it's
enough that when that money goes

556
00:32:18,760 --> 00:32:22,680
away, that does have an impact
on, you know, the quality of

557
00:32:22,680 --> 00:32:25,200
reporting.
And I, and I think though that

558
00:32:25,560 --> 00:32:30,160
they're really, I have a problem
with, with publicly funded news

559
00:32:30,160 --> 00:32:33,760
organisations because if you've
got a dictator, that then means

560
00:32:33,760 --> 00:32:38,360
the information is curated by
somebody who really has an axe

561
00:32:38,360 --> 00:32:40,000
to grind.
I don't.

562
00:32:40,000 --> 00:32:43,720
I don't see public as
necessarily the same as as

563
00:32:43,720 --> 00:32:45,720
unbiased.
I mean, yeah, that's that's a

564
00:32:45,720 --> 00:32:49,000
fair fairpoint.
I guess the flip side of that is

565
00:32:49,000 --> 00:32:53,360
that hypothetically, let's say
Donald Trump destroys it.

566
00:32:53,360 --> 00:32:56,480
Couldn't someone just reinstate
it in a few more years?

567
00:32:56,680 --> 00:33:00,160
Sure, absolutely.
And you know there was a time in

568
00:33:00,160 --> 00:33:03,440
the US where they had the
fairness doctrine and equal time

569
00:33:03,440 --> 00:33:06,160
provisions.
It got to be very difficult to

570
00:33:06,160 --> 00:33:10,520
administer and and a lot of
broadcasting organisations got

571
00:33:10,520 --> 00:33:12,320
their licences from the
government.

572
00:33:12,320 --> 00:33:16,160
So in order to get their licence
renewed they needed to play it

573
00:33:16,160 --> 00:33:18,320
straight and to be fair to all
sides.

574
00:33:18,600 --> 00:33:23,600
And when those particular rules
went away, it was the Wild West.

575
00:33:23,600 --> 00:33:29,000
It it gave rise to talk radio,
Rush Limbaugh, you know, people

576
00:33:29,000 --> 00:33:32,440
who you know, who didn't have
editors, they could spew

577
00:33:32,440 --> 00:33:34,680
whatever they wanted.
I think it's a pendulum.

578
00:33:34,680 --> 00:33:38,040
You know, we talked about truth.
We talked about objectivity.

579
00:33:38,360 --> 00:33:43,920
I think on a fundamental level,
people really do appreciate

580
00:33:43,920 --> 00:33:46,040
truth.
We don't like to be lied to,

581
00:33:46,560 --> 00:33:49,200
certainly don't like to be lied
to in relationships.

582
00:33:49,680 --> 00:33:52,840
And that's really what we're
talking about, a relationship

583
00:33:52,840 --> 00:33:56,560
between, you know, the people
and a president, the people and

584
00:33:56,560 --> 00:33:58,800
their government.
So I think that on some

585
00:33:58,800 --> 00:34:03,760
fundamental level there still is
a desirable for reliable

586
00:34:03,920 --> 00:34:06,840
information.
Yeah, 100%.

587
00:34:06,840 --> 00:34:10,440
And I think that that's one of
the benefits that the access to

588
00:34:10,440 --> 00:34:14,080
finding that truth for anybody
that is interested is you can

589
00:34:14,080 --> 00:34:21,199
just see how much these stuff is
kind of curated and you know,

590
00:34:21,199 --> 00:34:24,320
it's it's gives you a little bit
that doesn't give you the full

591
00:34:24,320 --> 00:34:26,920
story that you can find.
So you can obviously see what

592
00:34:26,920 --> 00:34:30,159
anyone's agenda and how much
anybody is is lying to you.

593
00:34:31,639 --> 00:34:35,040
Do you think people still value
journalism as a public good even

594
00:34:35,040 --> 00:34:36,560
if they're not willing to pay
for it?

595
00:34:36,560 --> 00:34:41,080
And does the amount they value
it, You know, like I, I value it

596
00:34:41,080 --> 00:34:44,159
quite a lot and I, I read a wide
mix of things.

597
00:34:44,159 --> 00:34:46,280
But do you think for a lot of
people that the amount they

598
00:34:46,280 --> 00:34:50,560
value it depends on how much it
aligns to their own, I guess

599
00:34:50,560 --> 00:34:54,239
mindset?
Yeah, because look, going back

600
00:34:54,239 --> 00:34:58,200
to when this country, the US,
was founded, you know, they

601
00:34:58,200 --> 00:35:01,000
didn't have anything called
objective journalism.

602
00:35:01,320 --> 00:35:04,040
You know, there were a lot of
newspapers, but each newspaper

603
00:35:04,040 --> 00:35:07,720
was spouting a particular
political position.

604
00:35:08,160 --> 00:35:12,120
And it wasn't until probably the
1920s, maybe 100 years ago,

605
00:35:12,360 --> 00:35:17,920
where the concept of objectivity
even entered journalism, entered

606
00:35:17,920 --> 00:35:21,640
the public sphere.
And, and so that was then

607
00:35:21,640 --> 00:35:25,080
curated and it became more
sophisticated.

608
00:35:25,360 --> 00:35:28,720
It was always controversial.
I think the Vietnam War was, was

609
00:35:28,720 --> 00:35:32,760
probably a perfect example of
that because, you know, Nixon

610
00:35:32,760 --> 00:35:36,360
was prosecuting a war and lying
about.

611
00:35:36,480 --> 00:35:40,200
And, and it wasn't just Nixon,
It was every president up until

612
00:35:40,200 --> 00:35:43,160
Nixon and beyond, you know,
lying about it.

613
00:35:43,440 --> 00:35:47,560
And during the Vietnam War, the
reporters that were actually in

614
00:35:47,560 --> 00:35:50,880
Vietnam covering it, they go in
the field with the troops and

615
00:35:51,160 --> 00:35:54,480
then they'd come back and there
would be a briefing in Saigon

616
00:35:54,480 --> 00:35:57,640
from the military leaders.
And they called the briefing the

617
00:35:57,640 --> 00:36:02,120
5:00 follies because there was
no, there was no coherence,

618
00:36:02,120 --> 00:36:05,800
because there was no, there was
no comparison to the lies that

619
00:36:05,800 --> 00:36:08,800
were told from the podium and
what these guys were seeing in

620
00:36:08,800 --> 00:36:11,640
the field.
But Nixon's vice President,

621
00:36:12,320 --> 00:36:15,640
Spiro Agnew was going around the
country calling journalists

622
00:36:15,880 --> 00:36:21,080
nattering nabobs of negativism
and and a feet core of impudent

623
00:36:21,080 --> 00:36:25,240
snobs, which is just another
sophisticated way of saying fake

624
00:36:25,240 --> 00:36:29,800
news.
So, you know, I I think we are

625
00:36:30,000 --> 00:36:33,720
fooling ourselves if we really
think that this is going to be

626
00:36:33,720 --> 00:36:36,520
solved.
There is no matter what the

627
00:36:36,520 --> 00:36:41,920
technology is, as long as people
are as involved, there is going

628
00:36:41,920 --> 00:36:45,520
to be funny business.
And I just, I sound like a

629
00:36:45,520 --> 00:36:48,960
broken record, but it just means
that those of us who care about

630
00:36:48,960 --> 00:36:53,400
the truth need to be discerning
and need to do what we can to

631
00:36:53,400 --> 00:36:56,520
let others know that not
everything you see on the

632
00:36:56,520 --> 00:36:59,480
Internet is trustworthy.
I think we all know that.

633
00:36:59,720 --> 00:37:03,360
But I think that politics has
gotten to be a blood sport.

634
00:37:03,600 --> 00:37:07,000
It's all about power.
It's all about power and

635
00:37:07,000 --> 00:37:10,520
winning.
And truth doesn't matter anymore

636
00:37:10,520 --> 00:37:12,880
because the bottom line is being
in control.

637
00:37:13,200 --> 00:37:16,320
And that's so sad.
It's really sad.

638
00:37:16,440 --> 00:37:19,880
Yeah, look that that again, 100%
agree with that.

639
00:37:19,960 --> 00:37:22,840
And I'm only speaking in an
Australian context.

640
00:37:22,840 --> 00:37:26,520
And from what I can see from the
US, it seems just kind of on

641
00:37:26,520 --> 00:37:32,360
another planet in that kind of
sense that even here it has kind

642
00:37:32,360 --> 00:37:36,680
of broken down to a team sport
to varying degrees.

643
00:37:36,680 --> 00:37:40,200
You know, it's kind of and just
that that hypocrisy and almost

644
00:37:40,200 --> 00:37:44,880
not holding anybody to anything
that they say of or promise.

645
00:37:44,880 --> 00:37:47,280
Yeah.
Do you see any spillover from

646
00:37:47,280 --> 00:37:50,560
the US where you know our
craziness is starting to seep

647
00:37:50,560 --> 00:37:51,840
into your politics?
Really.

648
00:37:51,840 --> 00:37:53,120
Of course.
Yeah, yeah, yeah.

649
00:37:53,120 --> 00:37:57,800
Look, I mean, rightly wrongly,
good or bad, like obviously the

650
00:37:57,840 --> 00:38:01,480
US is probably one of the, if
not the most influential culture

651
00:38:01,600 --> 00:38:05,360
in in the world, right.
And I think that being that

652
00:38:05,480 --> 00:38:09,480
highly visible, let's say leader
of the of the Western world that

653
00:38:09,480 --> 00:38:12,680
your political, cultural,
whatever it is, issues

654
00:38:13,160 --> 00:38:15,760
definitely bleed over into other
Western nations.

655
00:38:15,760 --> 00:38:20,080
And I think one of the, the
biggest ones is it's almost like

656
00:38:20,080 --> 00:38:22,520
we have bootleg versions like
you've got over there, you've

657
00:38:22,520 --> 00:38:24,760
got mega, right?
And then someone will, will

658
00:38:24,760 --> 00:38:28,600
repurpose that here and, and
it's make Australia great again.

659
00:38:28,600 --> 00:38:32,600
And, you know, similar kind of
sentiments, obviously, like

660
00:38:32,680 --> 00:38:35,960
Australia put Australians first
and, and, you know, same kind of

661
00:38:35,960 --> 00:38:37,480
thing.
And I think that it's not a, a

662
00:38:38,600 --> 00:38:44,280
uniquely Trumpian thing.
I think that it's just very easy

663
00:38:44,280 --> 00:38:49,640
to spread around so much shared
I guess culture and values in

664
00:38:49,640 --> 00:38:51,840
general.
Does that then mean that?

665
00:38:52,320 --> 00:38:56,680
You don't really have any other
places to go for reliable

666
00:38:56,680 --> 00:39:02,200
information because the Internet
has been so polluted entirely.

667
00:39:02,480 --> 00:39:07,120
Look, I would I would say.
That look and I could be

668
00:39:07,120 --> 00:39:09,560
completely wrong here, I'll just
talk from my perspective.

669
00:39:09,560 --> 00:39:14,200
I feel that, as I said, the US
being on another level with all

670
00:39:14,200 --> 00:39:18,000
of this stuff, it again comes
down to that like, I think here

671
00:39:18,440 --> 00:39:23,800
people aren't, from what I can
tell, as supremely emotionally

672
00:39:23,800 --> 00:39:26,600
invested in politics as as what
they are in the US.

673
00:39:26,600 --> 00:39:28,400
Like, you know, don't get me
wrong, we've got kind of

674
00:39:28,400 --> 00:39:31,920
politics super fans here, but
they seem to be in in far

675
00:39:31,920 --> 00:39:34,840
smaller numbers.
And I think that that what it's

676
00:39:34,840 --> 00:39:39,240
like here is most people
generally have this perception

677
00:39:39,240 --> 00:39:42,840
that the political class is just
completely mediocre.

678
00:39:43,120 --> 00:39:47,600
And so that doesn't have this
weight to the normal person.

679
00:39:47,600 --> 00:39:51,320
Like normal people are kind of
somewhat detached from it.

680
00:39:51,320 --> 00:39:53,840
And I think that that's a
benefit that we can enjoy.

681
00:39:55,360 --> 00:39:56,640
How did it get that way?
How do?

682
00:39:57,880 --> 00:40:00,000
Just how can we?
How can we learn from you?

683
00:40:00,600 --> 00:40:04,240
Look, I think.
I think find the least

684
00:40:04,240 --> 00:40:08,400
charismatic people you can find,
the least interesting people you

685
00:40:08,400 --> 00:40:13,720
can find and you know, like
straight down the middle

686
00:40:13,720 --> 00:40:17,160
inoffensive never rocked like
they're, they're careerists,

687
00:40:17,160 --> 00:40:18,800
right.
And look, I'm and again, we,

688
00:40:18,880 --> 00:40:21,120
we'll go off on a tangent, but
it is an interesting one.

689
00:40:21,560 --> 00:40:24,440
But it's like, you know, we know
that politicians are careerists,

690
00:40:24,440 --> 00:40:26,720
right?
Like what a what a job like get

691
00:40:26,720 --> 00:40:29,120
on that gravy train and you can
just ride it.

692
00:40:29,120 --> 00:40:32,800
And it's the same here, like,
you know, except here there's

693
00:40:33,960 --> 00:40:37,280
the salaries are probably far
less than what you would get in

694
00:40:37,280 --> 00:40:41,600
the US Politics here is kind of
got that control lever, but it

695
00:40:41,600 --> 00:40:46,480
doesn't have the cultural cache
that I guess, you know, like

696
00:40:46,480 --> 00:40:50,920
during election, someone might
put a a core flute or a sign in

697
00:40:50,920 --> 00:40:54,040
their yard or on their fence if
their local candidate.

698
00:40:54,040 --> 00:40:57,720
But it's like there's they've
never videos of someone driving

699
00:40:57,720 --> 00:41:00,480
their car through someone's yard
to smash them down.

700
00:41:00,480 --> 00:41:03,480
And fights breaking out and
things like that.

701
00:41:03,480 --> 00:41:05,320
It's just, it's just not like
that.

702
00:41:05,320 --> 00:41:08,960
And I think that at the end of
the day for the vast majority

703
00:41:08,960 --> 00:41:12,160
here, we are able to kind of
interact with each other.

704
00:41:12,160 --> 00:41:16,400
It does it hasn't broken society
tribally is is what I can

705
00:41:16,400 --> 00:41:19,760
observe.
But I think on on that while we

706
00:41:19,760 --> 00:41:23,920
were talking around, you know,
public broadcasters and and the

707
00:41:23,920 --> 00:41:27,920
search for truth, like, I think
one healthy thing on that here

708
00:41:27,920 --> 00:41:32,440
is this inherent cynicism, like
any, any claim that you might

709
00:41:32,440 --> 00:41:34,600
see from the from the
government, so many people are

710
00:41:34,600 --> 00:41:37,200
just like, that's bullshit.
You know, I'm gonna try and

711
00:41:37,200 --> 00:41:39,240
look, I'm gonna find why that's
wrong.

712
00:41:40,280 --> 00:41:42,360
So that's I guess that's one
benefit.

713
00:41:42,360 --> 00:41:45,600
And like I said, there is of
course, sycophants and

714
00:41:45,600 --> 00:41:48,440
mouthpieces, but I don't know,
it just doesn't have the

715
00:41:48,440 --> 00:41:51,320
cultural weight here.
I think that it might in the USI

716
00:41:52,280 --> 00:41:55,440
think though that.
Roughly the US is broken into

717
00:41:55,440 --> 00:41:57,680
thirds.
You know, there's the extreme

718
00:41:57,680 --> 00:42:00,680
right, the extreme left, and
then I think that there's the

719
00:42:00,680 --> 00:42:04,600
middle that really isn't as
tuned in anymore.

720
00:42:04,600 --> 00:42:06,880
There's sort of fatigue about
it.

721
00:42:06,880 --> 00:42:11,080
And I and I think that's where
I'm hopeful that those people,

722
00:42:11,440 --> 00:42:14,080
you know, might actually, you
know, swing it one way or the

723
00:42:14,160 --> 00:42:18,960
other if they're not persuadable
by 1 extreme or the other.

724
00:42:19,640 --> 00:42:20,840
Yeah, absolutely.
And look that.

725
00:42:20,840 --> 00:42:23,920
That pretty much sounds like
like here, most people are in

726
00:42:23,920 --> 00:42:26,960
that kind of grey area.
And I think that the, the, the

727
00:42:26,960 --> 00:42:29,360
good thing about that is
obviously being able to find

728
00:42:29,360 --> 00:42:32,920
common cause with people that
aren't on your team as such.

729
00:42:32,920 --> 00:42:36,160
And then that's that shared
humanity, which obviously can

730
00:42:36,160 --> 00:42:38,880
help hopefully move, move
society along.

731
00:42:39,080 --> 00:42:41,080
I think we're, I think you're
really on to something.

732
00:42:41,080 --> 00:42:44,520
I feel very strongly about that,
that common human ground.

733
00:42:44,760 --> 00:42:45,360
Yeah.
Look.

734
00:42:45,360 --> 00:42:47,840
Fingers crossed.
But while we're on a positive

735
00:42:47,840 --> 00:42:51,920
note, what would you say one
thing that the introduction of

736
00:42:51,920 --> 00:42:55,080
the Internet and and technology
has enabled in the world of

737
00:42:55,360 --> 00:42:58,040
journalism and writing that's
made you more hopeful about the

738
00:42:58,040 --> 00:42:59,800
future of it?
Being able to have this

739
00:42:59,800 --> 00:43:02,680
conversation I.
Think is a perfect example I

740
00:43:02,680 --> 00:43:05,880
mean you're concerned about what
the Internet is doing to

741
00:43:06,120 --> 00:43:09,880
society, the world, and we're
talking about that.

742
00:43:09,880 --> 00:43:12,040
We are half a world away talking
about it.

743
00:43:12,360 --> 00:43:14,600
I think that is what gives me
hope.

744
00:43:14,840 --> 00:43:16,440
Awesome.
On that point then we.

745
00:43:16,680 --> 00:43:19,960
Need those people to believe in
and and deliver that hope moving

746
00:43:19,960 --> 00:43:22,160
forward from from within the
industry.

747
00:43:22,480 --> 00:43:25,320
What would you say in your
opinion is the most important

748
00:43:25,320 --> 00:43:28,520
skill for a young writer,
whatever form of writing that

749
00:43:28,520 --> 00:43:30,880
they're, you know, looking to
pursue?

750
00:43:31,600 --> 00:43:34,640
The most important skill for
them to have today that was not

751
00:43:34,640 --> 00:43:39,040
needed, say, 20 years ago.
Boy I I hadn't.

752
00:43:39,040 --> 00:43:41,800
Thought about that one because I
mean, you know, the obvious

753
00:43:41,800 --> 00:43:46,920
skill is just being able to have
a vocabulary and be able to put

754
00:43:46,920 --> 00:43:51,840
into words succinctly and
quickly whatever it is that

755
00:43:51,840 --> 00:43:56,560
you're trying to communicate.
I think here's here's one

756
00:43:56,560 --> 00:44:00,360
thought just off the top of my
head, and that is it's almost

757
00:44:00,360 --> 00:44:05,200
the antithesis of the Blizzard
of information that we get.

758
00:44:05,400 --> 00:44:11,000
It would seem to me that the
ultimate is to be able to say

759
00:44:11,000 --> 00:44:15,760
something quickly and succinctly
and speak in sound bites, get

760
00:44:15,760 --> 00:44:18,840
the point across without
belabouring it.

761
00:44:19,520 --> 00:44:22,440
And now I'm going to stop.
No, that's that's an

762
00:44:22,440 --> 00:44:23,760
interesting.
Point there around especially

763
00:44:23,760 --> 00:44:27,560
since we've been talking about
people reading that headline or

764
00:44:27,560 --> 00:44:30,920
just the top level of it.
So I wonder if that's the skill

765
00:44:30,920 --> 00:44:36,680
to develop is figuring out a way
to deliver a real truthful

766
00:44:37,320 --> 00:44:40,080
headline or or sound bite that
is actually going to deliver

767
00:44:40,480 --> 00:44:44,560
that truth rather than just that
click and that intrigue to then

768
00:44:44,560 --> 00:44:47,000
deliver.
I don't know something else with

769
00:44:47,000 --> 00:44:50,760
a different different agenda A
lot to think about.

770
00:44:51,520 --> 00:44:53,440
I know a lot to.
Think about.

771
00:44:54,640 --> 00:44:58,400
Just to finish up then, what
advice would you give to anybody

772
00:44:58,400 --> 00:45:02,760
either you know currently within
the industry or thinking of

773
00:45:02,760 --> 00:45:06,080
getting into it sometime in the
near future or or down the line,

774
00:45:06,600 --> 00:45:10,400
what how would you advise them
to not only navigate the current

775
00:45:10,960 --> 00:45:14,160
landscape and world of
journalism and writing, but also

776
00:45:14,160 --> 00:45:16,200
help future proof themselves as
well?

777
00:45:16,320 --> 00:45:18,360
Be curious.
I think that's.

778
00:45:18,360 --> 00:45:21,680
Fundamental I I don't think that
that will ever go away.

779
00:45:21,680 --> 00:45:27,480
And not only be be curious, but
be assertive about your

780
00:45:27,480 --> 00:45:30,440
curiosity.
Don't be afraid to ask questions

781
00:45:30,560 --> 00:45:34,880
and and why is a wonderful
question. 5 year olds, you know,

782
00:45:34,880 --> 00:45:36,720
get it, You know, it's time for
bed.

783
00:45:36,720 --> 00:45:39,840
Why?
Because I said so why and so on.

784
00:45:41,320 --> 00:45:46,720
So I think curiosity will never
go out of fashion and and asking

785
00:45:46,720 --> 00:45:49,680
The Who, what, where, when, why
and how questions.

786
00:45:50,760 --> 00:45:54,000
Yeah, I think stay curious.
Is a great piece of advice not

787
00:45:54,000 --> 00:45:56,920
just for anybody in that world,
but obviously anybody in the

788
00:45:57,000 --> 00:46:00,560
world itself.
That's how people adopt that as

789
00:46:00,560 --> 00:46:03,080
it's kind of a mindset and we
can maybe get through this

790
00:46:03,080 --> 00:46:08,240
quagmire of mess using the
Internet to to help find those

791
00:46:08,280 --> 00:46:11,280
pieces of truth.
I'm hopeful for no apparent.

792
00:46:11,280 --> 00:46:12,880
Reason.
That's good.

793
00:46:12,880 --> 00:46:13,880
Me too.
Me too.

794
00:46:14,720 --> 00:46:18,000
Thanks so much for that, John.
What have you got coming up on

795
00:46:18,000 --> 00:46:20,000
the horizon?
Where can people follow what

796
00:46:20,000 --> 00:46:22,640
you're up to?
Probably the best thing is my

797
00:46:22,640 --> 00:46:24,920
website.
Which is myname.com,

798
00:46:24,920 --> 00:46:34,240
johndidakis.com, JOHND as in dog
Ed as in dog Akis as in

799
00:46:34,240 --> 00:46:37,760
samjohndidakis.com.
And I think that there are

800
00:46:37,760 --> 00:46:40,160
actually some Didakis is living
in Australia.

801
00:46:40,400 --> 00:46:41,600
Oh, really?
I have to look that up.

802
00:46:41,640 --> 00:46:44,320
Yeah, that.
Migrated from Lofka.

803
00:46:44,520 --> 00:46:47,840
In in Greece on the Peloponnese,
I was just there where my

804
00:46:47,840 --> 00:46:49,640
grandpa and great grandpa were
born.

805
00:46:50,040 --> 00:46:52,000
Yeah, right.
Yeah, we do have a decent.

806
00:46:52,000 --> 00:46:54,480
Greek population here, so it
wouldn't wouldn't surprise me.

807
00:46:54,480 --> 00:46:57,360
Yeah, right.
Pasiotis is another.

808
00:46:57,680 --> 00:47:00,480
Greek family that moved to
Australia.

809
00:47:01,160 --> 00:47:04,000
So yeah, my website is probably
the best place to do.

810
00:47:04,240 --> 00:47:08,000
I'm doing more public speaking
now on helping people use

811
00:47:08,160 --> 00:47:10,080
writing as a way to heal from
grief.

812
00:47:10,080 --> 00:47:14,000
I'm working on my 7th novel.
I have a short story that's

813
00:47:14,160 --> 00:47:17,480
that's with an editor and I've
written a memoir that's with a

814
00:47:17,480 --> 00:47:20,400
publisher and they're deciding
whether to publish it.

815
00:47:20,400 --> 00:47:24,240
So I've got, you know, plus I'm
teaching classes online and

816
00:47:24,240 --> 00:47:27,480
because it's online on the
Internet, you could even take

817
00:47:27,480 --> 00:47:30,040
one of my classes even though
you're in Australia.

818
00:47:30,920 --> 00:47:33,360
And where can people find find?
Those it's just at your website.

819
00:47:33,360 --> 00:47:35,560
Go to my website, go to upcoming
events.

820
00:47:35,560 --> 00:47:38,440
And you'll find it awesome,
John, thank.

821
00:47:38,440 --> 00:47:40,360
You so much.
Thank you, Gareth, it was

822
00:47:40,360 --> 00:47:41,360
wonderful talking.
To you.

823
00:47:41,360 --> 00:47:44,560
Thank you for more info on what
we've discussed today.

824
00:47:44,560 --> 00:47:46,640
Check out the show notes if you
enjoyed this one.

825
00:47:46,680 --> 00:47:49,120
You can subscribe to Ruined by
the Internet on your favourite

826
00:47:49,120 --> 00:47:52,240
podcast app and help spread the
word by sharing this episode or

827
00:47:52,240 --> 00:47:54,920
leaving a review.
I'm Gareth King, see you next

828
00:47:54,920 --> 00:47:55,160
time.
John DeDakis Profile Photo

John DeDakis

Award-winning novelist John DeDakis (Pronounced: dee-DAY-kiss) is a former editor on CNN's "The Situation Room with Wolf Blitzer." DeDakis is the author of six novels in the Lark Chadwick mystery-suspense-thriller series. In his most recent book, Enemies Domestic, Lark is a White House press secretary forced to make her extremely personal abort-or-not-to-abort decision in a highly toxic and polarized political fishbowl—all while dealing with an attack on the presidency. DeDakis, a former White House correspondent, is a writing coach, manuscript editor, podcaster, and regularly teaches novel writing at literary centers and writers’ conferences nationwide. Website: www.johndedakis.com