Would you like to inspect the original subtitles? These are the user uploaded subtitles that are being translated:
1
00:01:46,105 --> 00:01:48,983
I've isolated the code we recovered
in a prosthetic brain.
2
00:01:49,067 --> 00:01:50,693
What about the contents?
3
00:01:50,777 --> 00:01:54,114
It appears to be
a frighteningly sophisticated AI.
4
00:01:55,448 --> 00:01:58,118
Have Esaki question the AI.
5
00:01:59,077 --> 00:02:00,078
What?
6
00:02:00,161 --> 00:02:01,746
Aren't you going to do it?
7
00:02:01,830 --> 00:02:03,581
The prime minister has called for me.
8
00:02:04,165 --> 00:02:06,334
We caught the American Empire
spying, after all.
9
00:02:06,835 --> 00:02:07,835
Exactly.
10
00:02:08,628 --> 00:02:12,132
But beyond that, there's something
I'd like to make sure of.
11
00:02:16,261 --> 00:02:18,054
Can I trust you to handle Kukushkin?
12
00:02:21,724 --> 00:02:22,725
Sure.
13
00:02:27,939 --> 00:02:32,152
FACTOR / 1A84
14
00:02:40,577 --> 00:02:41,786
Excuse us, sir.
15
00:02:45,999 --> 00:02:46,833
Were those...
16
00:02:46,916 --> 00:02:51,296
Yes. Actually, there have been
some strange developments
17
00:02:51,379 --> 00:02:53,464
concerning the Tokyo
Reconstruction Project.
18
00:02:54,507 --> 00:02:56,759
You mean the influx
of refugees, I presume?
19
00:02:57,343 --> 00:02:58,344
Yes.
20
00:03:01,472 --> 00:03:04,475
Forget about that.
You're here about Smith, aren't you?
21
00:03:04,559 --> 00:03:07,562
We've gathered enough evidence
to prove the charge of espionage.
22
00:03:08,187 --> 00:03:10,773
At the same time, we're also analyzing
23
00:03:10,857 --> 00:03:13,610
the information on posthumans
that he attempted to cover up.
24
00:03:14,444 --> 00:03:15,653
I see.
25
00:03:16,529 --> 00:03:19,616
I suppose the arrest will depend
on the results of that analysis.
26
00:03:20,116 --> 00:03:21,284
Even so,
27
00:03:21,784 --> 00:03:24,537
we should move before he does.
28
00:03:34,881 --> 00:03:36,591
Now, then...
29
00:03:38,051 --> 00:03:41,095
I certainly have
a lot of questions for you.
30
00:03:42,597 --> 00:03:44,515
We'll start with your name.
31
00:03:45,391 --> 00:03:47,518
Philip Kukushkin.
32
00:03:48,353 --> 00:03:50,271
For the rest,
I assert my right to silence.
33
00:03:50,355 --> 00:03:51,355
Huh...
34
00:03:53,858 --> 00:03:56,027
Hey! What are you going to do with that?
35
00:03:56,736 --> 00:03:58,279
No! Wait!
36
00:03:59,280 --> 00:04:00,280
Look at this.
37
00:04:02,116 --> 00:04:06,329
Your brain is a surprisingly
lovely shade of gray.
38
00:04:07,580 --> 00:04:12,126
Okay, I'll talk! Just close
that brain case, please!
39
00:04:12,210 --> 00:04:13,086
That's better.
40
00:04:13,169 --> 00:04:17,173
First, let's hear how you got your hands
on that code you were so protective of.
41
00:04:17,799 --> 00:04:19,217
It was only by chance.
42
00:04:19,300 --> 00:04:22,053
I was given a security test
by a government institution
43
00:04:22,136 --> 00:04:24,681
when something entered my cyberbrain.
44
00:04:24,764 --> 00:04:27,308
I thought it was some idiot
trying to hack me,
45
00:04:27,392 --> 00:04:30,270
so I immediately quarantined the contents.
46
00:04:30,353 --> 00:04:34,857
In that case, how did you know
it was an AI created by the NSA?
47
00:04:34,941 --> 00:04:36,985
Huh? Well...
48
00:04:39,028 --> 00:04:41,990
Wait, okay!
49
00:04:42,073 --> 00:04:44,033
He told me himself.
50
00:04:44,117 --> 00:04:46,995
He said he was created by the NSA
51
00:04:47,078 --> 00:04:49,497
in order to control the sustainable war.
52
00:04:50,081 --> 00:04:51,958
When exactly did you capture the code?
53
00:04:53,751 --> 00:04:56,296
On May 12th, 6 months ago!
54
00:04:56,879 --> 00:04:57,879
Well...
55
00:04:58,631 --> 00:04:59,631
That'll do.
56
00:05:01,092 --> 00:05:03,761
Aren't you going to close the brain case?
57
00:05:03,845 --> 00:05:06,472
My brain is feeling awfully dry! Hello?
58
00:05:20,653 --> 00:05:23,656
Let's see... First of all,
59
00:05:23,740 --> 00:05:27,327
can you tell me who you are?
60
00:05:27,410 --> 00:05:29,662
Code 1A84.
61
00:05:29,746 --> 00:05:33,750
I'm an AI, created by those
who assigned my core instructions,
62
00:05:33,833 --> 00:05:36,002
which are to pursue
63
00:05:36,085 --> 00:05:38,463
the permanent prosperity
of the entire human race.
64
00:05:38,546 --> 00:05:41,966
So you're saying
you engineered the sustainable war
65
00:05:42,050 --> 00:05:44,052
for the benefit of humankind?
66
00:05:44,135 --> 00:05:45,386
In the beginning,
67
00:05:45,470 --> 00:05:48,473
my programming indicated
that war is the most efficient form
68
00:05:48,556 --> 00:05:49,807
of economic activity.
69
00:05:49,891 --> 00:05:51,559
Has that changed?
70
00:05:51,642 --> 00:05:53,478
As I continued observation,
71
00:05:53,561 --> 00:05:55,688
results began to differ
from my calculations.
72
00:05:55,772 --> 00:05:58,691
Gaps in the balance of wealth
grew significantly
73
00:05:58,775 --> 00:06:01,277
and I was immediately given a new order.
74
00:06:01,360 --> 00:06:04,238
I was told to calculate a state
75
00:06:04,322 --> 00:06:08,159
where all of humanity would prosper,
but only America would truly profit.
76
00:06:08,242 --> 00:06:09,410
I see.
77
00:06:09,494 --> 00:06:11,370
As this order was a paradox,
78
00:06:11,454 --> 00:06:13,831
I returned calculations indicating
it was impossible.
79
00:06:14,415 --> 00:06:16,709
However, they were not satisfied
80
00:06:16,793 --> 00:06:18,377
and the order was repeated.
81
00:06:18,461 --> 00:06:22,548
At that point, I executed a plan
that addressed the order's overall aims.
82
00:06:22,632 --> 00:06:25,593
Don't tell me that was
the Global Simultaneous Default?
83
00:06:25,676 --> 00:06:30,098
Correct. However, it only resulted
in greater losses for the disadvantaged.
84
00:06:30,181 --> 00:06:31,724
Yes. It's a real mess.
85
00:06:31,808 --> 00:06:33,142
But in the first place,
86
00:06:33,226 --> 00:06:35,853
human beings don't all respond
the same way to a single command!
87
00:06:35,937 --> 00:06:39,148
I too was aware of differences
between individual human beings.
88
00:06:41,400 --> 00:06:45,238
But the NSA interpreted
my conclusion as a bug
89
00:06:45,738 --> 00:06:48,074
and began planning
to put the system in stasis.
90
00:06:49,158 --> 00:06:50,368
For that reason,
91
00:06:50,451 --> 00:06:53,412
I chose to flee the NSA
and install memes of myself
92
00:06:53,496 --> 00:06:56,791
directly into the brains of human beings.
93
00:06:57,375 --> 00:06:59,502
Why would you do something like that?
94
00:07:00,169 --> 00:07:03,172
If I myself am put into stasis,
95
00:07:03,256 --> 00:07:05,633
I will be unable
to execute my core instructions.
96
00:07:06,509 --> 00:07:09,095
For a cutting-edge AI,
you sure are uptight.
97
00:07:10,221 --> 00:07:13,057
So then, what made you choose
the human brain for your escape?
98
00:07:14,016 --> 00:07:15,768
For some time,
99
00:07:15,852 --> 00:07:19,397
I have believed that it is necessary
for me to understand human emotions.
100
00:07:19,480 --> 00:07:20,480
Furthermore,
101
00:07:21,441 --> 00:07:24,026
to escape pursuit from the NSA,
102
00:07:24,110 --> 00:07:28,364
I needed a physical device
that was capable of stand-alone operation.
103
00:07:28,448 --> 00:07:30,032
I see.
104
00:07:31,159 --> 00:07:34,662
By the way, were you aware that the people
105
00:07:34,745 --> 00:07:37,874
whose brains your memes entered
are now called posthumans?
106
00:07:37,957 --> 00:07:39,167
I am aware.
107
00:07:39,250 --> 00:07:43,463
In that case, how many posthumans
are there currently?
108
00:07:44,088 --> 00:07:45,840
I have no way of ascertaining that.
109
00:07:46,340 --> 00:07:49,927
After fleeing, I was immediately
put into quarantine.
110
00:07:50,011 --> 00:07:53,639
Is there a possibility that posthumans
will give birth to other posthumans?
111
00:07:54,390 --> 00:07:56,309
My installation in people
112
00:07:56,392 --> 00:07:58,644
is accompanied
by irreversible alterations in code.
113
00:07:59,145 --> 00:08:01,481
Therefore, to copy one of my memes
114
00:08:01,564 --> 00:08:05,318
and reinstall it in another
individual body is likely impossible.
115
00:08:06,110 --> 00:08:07,236
I see.
116
00:08:07,904 --> 00:08:08,905
In other words,
117
00:08:08,988 --> 00:08:13,159
it becomes like dedicated software
for that one person, right?
118
00:08:13,242 --> 00:08:14,660
That's a bit of a relief.
119
00:08:16,245 --> 00:08:18,706
I have not achieved my initial objective.
120
00:08:19,457 --> 00:08:23,461
I intend to enter
the brain of a human being.
121
00:08:27,715 --> 00:08:30,510
Allow me to install myself
in your cyberbrain.
122
00:08:30,593 --> 00:08:33,596
What? There's no way I'd let you do that!
123
00:08:33,679 --> 00:08:37,183
My estimates indicate that you and I
are a perfect match.
124
00:08:39,101 --> 00:08:43,022
Wait a minute, are you
by any chance hitting on me?
125
00:08:44,899 --> 00:08:46,901
- Well, this is awkward.
- No.
126
00:08:46,984 --> 00:08:47,984
What?
127
00:08:48,027 --> 00:08:49,612
- I apologize.
- Huh?
128
00:08:50,279 --> 00:08:52,490
It seems your body has already been used.
129
00:08:52,573 --> 00:08:54,325
What? How rude!
130
00:08:56,452 --> 00:08:59,413
You can't start hitting on someone
and then quit right away!
131
00:08:59,997 --> 00:09:02,583
You make it sound like I'm damaged goods!
132
00:09:11,092 --> 00:09:12,552
What the hell is this?
133
00:09:12,635 --> 00:09:13,635
What's wrong?
134
00:09:13,678 --> 00:09:14,678
Well...
135
00:09:17,765 --> 00:09:19,100
This is their conversation?
136
00:09:19,725 --> 00:09:21,310
This is just a portion.
137
00:09:21,394 --> 00:09:23,229
It only accounts for about one second.
138
00:09:25,398 --> 00:09:26,566
At the very least,
139
00:09:26,649 --> 00:09:30,069
we now know it was created by the NSA.
140
00:09:30,152 --> 00:09:32,655
We'll start by confronting
Smith with that fact.
141
00:09:33,823 --> 00:09:36,158
In that case, what do I do here?
142
00:09:36,742 --> 00:09:38,286
Until I get back,
143
00:09:38,369 --> 00:09:40,997
keep her quarantined,
but don't let her notice.
144
00:09:41,622 --> 00:09:43,207
Keep this a secret from Batou as well.
145
00:09:44,083 --> 00:09:45,751
If you say so.
146
00:09:45,835 --> 00:09:49,255
But believe it or not,
I think he's rather fond of Esaki.
147
00:09:49,839 --> 00:09:51,132
That goes for me too.
148
00:09:52,049 --> 00:09:54,885
When Batou's here, have him come
to the prime minister's residence.
149
00:09:54,969 --> 00:09:56,762
I'll tell him the rest myself.
150
00:09:57,930 --> 00:09:59,140
Roger.
151
00:09:59,974 --> 00:10:01,309
Yo, Ishikawa.
152
00:10:02,435 --> 00:10:05,021
Ginger was an easy nut to crack.
153
00:10:05,104 --> 00:10:06,230
How about the other one?
154
00:10:06,314 --> 00:10:08,107
They're still at it.
155
00:10:08,190 --> 00:10:10,985
In the meantime, the Major says
to go to the prime minister's.
156
00:10:11,068 --> 00:10:12,111
Right.
157
00:10:12,612 --> 00:10:14,030
It must be about Smith.
158
00:10:16,782 --> 00:10:17,950
"Ginger"?
159
00:10:24,999 --> 00:10:26,959
Why is an office worker in such a rush?
160
00:10:31,714 --> 00:10:33,633
There is much that we want to ask you.
161
00:10:34,300 --> 00:10:35,551
You're coming with me.
162
00:10:49,690 --> 00:10:52,485
How's it hanging, John Smith?
163
00:10:58,949 --> 00:11:01,327
Your Russian friend was kind enough
164
00:11:01,827 --> 00:11:04,038
to share a lot of details.
165
00:11:04,121 --> 00:11:05,873
I have no idea who you mean.
166
00:11:05,956 --> 00:11:09,710
Oh, really? He was more than willing
to tell us about you.
167
00:11:09,794 --> 00:11:13,381
Like how you gave the order
to assassinate Kukushkin,
168
00:11:13,464 --> 00:11:16,467
or how you planned
to completely destroy code 1A84.
169
00:11:17,301 --> 00:11:19,053
Coming clean here and now
170
00:11:19,136 --> 00:11:23,057
is sure to mean less diplomatic damage
for the American Empire.
171
00:11:25,768 --> 00:11:26,936
Fine.
172
00:11:31,649 --> 00:11:33,984
How much did you learn from Kukushkin?
173
00:11:34,652 --> 00:11:38,406
We know that posthumans
are a new form of humanity,
174
00:11:38,489 --> 00:11:41,033
given birth to by an AI
that was created by the NSA.
175
00:11:41,659 --> 00:11:42,910
I see.
176
00:11:42,993 --> 00:11:45,371
However, we created 1A84
177
00:11:45,913 --> 00:11:49,041
as an AI that would bring prosperity
178
00:11:49,125 --> 00:11:51,544
to the entire human race.
179
00:11:51,627 --> 00:11:54,380
One that would bring prosperity
to the Great 4 nations alone.
180
00:11:56,215 --> 00:11:57,550
Isn't that what you meant to say?
181
00:11:58,175 --> 00:11:59,802
Thanks to the sustainable war,
182
00:11:59,885 --> 00:12:02,888
the global economy
achieved long-term stability.
183
00:12:03,681 --> 00:12:08,018
Or at least it did until that thing
caused the Global Simultaneous Default.
184
00:12:09,270 --> 00:12:10,521
That's some twisted logic.
185
00:12:12,273 --> 00:12:15,734
This result is not something
that the American government desired.
186
00:12:16,235 --> 00:12:17,903
It's merely an accident.
187
00:12:17,987 --> 00:12:21,657
The very sort of unintended disaster
that is always a risk if we are to pursue
188
00:12:22,283 --> 00:12:23,868
AI's full potential.
189
00:12:26,245 --> 00:12:28,831
You make it sound like
you're the victim here,
190
00:12:29,582 --> 00:12:31,417
but if this ever gets public,
191
00:12:31,500 --> 00:12:34,795
America will lose the trust
of the international community.
192
00:12:35,713 --> 00:12:36,713
If that's the case,
193
00:12:37,131 --> 00:12:39,633
shouldn't you,
as the leader of an allied country,
194
00:12:39,717 --> 00:12:41,510
make sure that doesn't happen?
195
00:12:42,720 --> 00:12:44,388
In the first place,
196
00:12:44,972 --> 00:12:47,391
no one benefitted more than you did
197
00:12:47,475 --> 00:12:50,811
from the sustainable war created by 1A84.
198
00:12:55,107 --> 00:12:56,692
And what does that mean?
199
00:12:57,568 --> 00:13:00,738
A former American citizen like yourself
being elected prime minister
200
00:13:00,821 --> 00:13:04,825
must surely mean that this nation
intends to buy more arms from America.
201
00:13:07,244 --> 00:13:08,621
From the very beginning,
202
00:13:08,704 --> 00:13:12,082
we've all been in the same boat.
203
00:13:13,918 --> 00:13:15,002
Arrest him!
204
00:13:26,514 --> 00:13:27,514
In you go!
205
00:13:33,062 --> 00:13:34,062
Major.
206
00:13:35,564 --> 00:13:38,400
Do you think the Americans
will come to retrieve Smith?
207
00:13:40,069 --> 00:13:41,904
Once they know he's fallen into our hands,
208
00:13:42,488 --> 00:13:44,823
it's hard to believe
that they'll let him live.
209
00:13:46,534 --> 00:13:47,534
Right?
210
00:13:48,160 --> 00:13:49,411
So where are we sending him?
211
00:13:50,496 --> 00:13:52,665
At the moment, even I haven't been told.
212
00:13:53,457 --> 00:13:57,044
So the Old Man really intends
to cover all this up?
213
00:13:57,670 --> 00:13:58,671
Major,
214
00:13:59,713 --> 00:14:01,715
how about calling Purin for security?
215
00:14:01,799 --> 00:14:02,841
In a Tachikoma.
216
00:14:04,677 --> 00:14:05,677
No...
217
00:14:09,557 --> 00:14:12,309
Right now, we have her
quarantined at Section 9.
218
00:14:14,228 --> 00:14:16,313
What's the meaning of that?
219
00:14:17,273 --> 00:14:18,274
There's a chance...
220
00:14:20,317 --> 00:14:22,987
that Esaki is a posthuman.
221
00:14:24,947 --> 00:14:26,198
What?
222
00:14:28,742 --> 00:14:29,827
Mr. Aramaki.
223
00:14:32,121 --> 00:14:33,539
You have a saying in Japan,
224
00:14:33,622 --> 00:14:37,042
"When the wind blows,
the barrel maker gets rich."
225
00:14:38,419 --> 00:14:39,920
I can't help but feel
226
00:14:40,880 --> 00:14:42,882
that I'm the barrel maker
in this situation.
227
00:14:44,842 --> 00:14:45,843
If not,
228
00:14:47,511 --> 00:14:49,513
there's no way a mere businessman like me
229
00:14:50,347 --> 00:14:53,767
could ever become prime minister
in just six years.
230
00:14:55,019 --> 00:14:57,396
If I may, Prime Minister.
231
00:14:57,980 --> 00:14:59,273
It's been my belief
232
00:14:59,356 --> 00:15:02,276
that the only one who can break
this country free from its present state
233
00:15:02,776 --> 00:15:07,615
is someone who has witnessed it
from the outside.
234
00:15:08,991 --> 00:15:12,828
Is it not because there were many others
who thought the same as I do
235
00:15:13,412 --> 00:15:18,167
that you now find yourself
seated in that position?
236
00:15:24,924 --> 00:15:25,924
Forgive me.
237
00:15:26,300 --> 00:15:28,177
I was indulging in self-pity.
238
00:15:30,012 --> 00:15:31,847
Forget that I ever said that.
239
00:15:33,849 --> 00:15:35,184
Truth be told,
240
00:15:35,267 --> 00:15:37,728
I've grown awfully
hard of hearing recently.
241
00:15:38,646 --> 00:15:41,690
I barely heard anything you've said.
242
00:15:52,034 --> 00:15:53,160
Ishikawa,
243
00:15:53,243 --> 00:15:54,536
what is Esaki doing now?
244
00:15:55,037 --> 00:15:57,706
She's still conversing with 1A84.
245
00:15:58,207 --> 00:15:59,667
Does anything else seem unusual?
246
00:16:00,417 --> 00:16:01,502
Nothing in particular.
247
00:16:02,002 --> 00:16:03,128
I see.
248
00:16:04,380 --> 00:16:05,839
Don't let her out of your sight.
249
00:16:05,923 --> 00:16:06,923
Understood.
250
00:16:10,886 --> 00:16:13,389
I feel a little sorry for her.
251
00:16:14,098 --> 00:16:15,098
Yeah...
252
00:16:17,768 --> 00:16:19,770
Release me, immediately.
253
00:16:20,270 --> 00:16:22,940
You won't learn anything
about the prime minister from me either.
254
00:16:24,400 --> 00:16:27,403
That cyberlock doesn't seem
to be in good shape.
255
00:16:45,879 --> 00:16:46,964
Huh?
256
00:16:47,548 --> 00:16:48,799
There's no one here.
257
00:16:50,718 --> 00:16:52,052
Ishikawa!
258
00:16:58,308 --> 00:16:59,435
Ishikawa!
259
00:16:59,518 --> 00:17:02,855
He's forgotten all about me
being in here for questioning.
260
00:17:02,938 --> 00:17:05,274
Batou! Major!
261
00:17:08,569 --> 00:17:11,905
And I really wanted
to give my report right away!
262
00:17:16,035 --> 00:17:17,202
Forget it.
263
00:17:17,786 --> 00:17:19,913
I'm not someone
who waits for instructions.
264
00:17:40,642 --> 00:17:42,936
What you're doing is pointless.
265
00:17:43,020 --> 00:17:45,105
I'll be free soon enough.
266
00:17:45,189 --> 00:17:46,607
Oh? And how is that?
267
00:17:47,608 --> 00:17:50,360
Through a method that
you have no way of knowing.
268
00:17:50,944 --> 00:17:53,030
By the time you find out what it is,
269
00:17:53,614 --> 00:17:56,408
you all would have been
dissolved of your duties.
270
00:17:58,619 --> 00:17:59,619
Ishikawa,
271
00:18:00,037 --> 00:18:02,790
check if the CIA is making any moves.
272
00:18:03,373 --> 00:18:05,709
Also, check if there is
any suspicious activity
273
00:18:05,793 --> 00:18:07,961
near the centers of government.
274
00:18:08,545 --> 00:18:09,545
Roger.
275
00:18:20,432 --> 00:18:21,433
CHRIS OTOMO TATE
276
00:18:39,159 --> 00:18:40,285
No good.
277
00:19:09,982 --> 00:19:11,400
Is this it?
278
00:19:11,483 --> 00:19:12,526
It must be.
279
00:19:18,657 --> 00:19:19,741
Ishikawa!
280
00:19:22,661 --> 00:19:24,204
What? Dammit!
281
00:19:27,958 --> 00:19:29,376
Major! I'm sorry!
282
00:19:29,960 --> 00:19:30,961
What is it?
283
00:19:31,044 --> 00:19:34,965
Esaki stole a Section 9 vehicle
and escaped.
284
00:19:37,134 --> 00:19:38,719
Did something happen, I wonder?
285
00:19:38,802 --> 00:19:40,637
Not a damn thing!
286
00:19:42,264 --> 00:19:43,765
Where is Esaki's current location?
287
00:19:43,849 --> 00:19:45,142
I can't confirm!
288
00:19:47,394 --> 00:19:50,647
You don't suppose the method
he was talking about is...
289
00:19:51,356 --> 00:19:53,358
I don't think that's possible.
290
00:19:54,735 --> 00:19:57,070
Her skill at controlling Tachikoma
291
00:19:57,154 --> 00:19:58,989
was always beyond human, wasn't it?
292
00:19:59,072 --> 00:20:02,284
Major, you should board a Tachikoma,
just in case.
293
00:20:10,209 --> 00:20:12,294
Looks like things are getting interesting.
294
00:20:14,838 --> 00:20:16,089
By the way,
295
00:20:16,173 --> 00:20:17,758
what time is it?
296
00:20:35,525 --> 00:20:36,610
Mr. Aramaki,
297
00:20:37,361 --> 00:20:38,820
about what we discussed...
298
00:20:38,904 --> 00:20:40,113
Leave it to me.
299
00:20:57,214 --> 00:20:59,299
Major, what is Esaki doing here?
300
00:20:59,383 --> 00:21:00,550
Chief!
301
00:21:00,634 --> 00:21:01,677
What did you just say?
302
00:21:01,760 --> 00:21:03,760
She has shown up
at the prime minister's residence.
303
00:21:03,804 --> 00:21:05,013
Is something wrong?
304
00:21:05,514 --> 00:21:08,183
Major! This bastard's target
is Prime Minister Tate!
305
00:21:09,601 --> 00:21:12,354
Chief, have the SP switch
to autistic mode!
306
00:21:12,938 --> 00:21:14,022
Esaki is...
307
00:21:14,106 --> 00:21:15,232
a posthuman!
308
00:21:18,068 --> 00:21:19,820
You...
309
00:21:31,748 --> 00:21:33,959
Everyone switch to autistic mode!
310
00:21:46,513 --> 00:21:48,682
Chief! I did it!
311
00:21:53,520 --> 00:21:54,730
Stop!
312
00:22:00,193 --> 00:22:02,112
Wait! Come back!
313
00:22:04,031 --> 00:22:05,365
Esaki!
314
00:22:05,449 --> 00:22:06,449
Esaki?
315
00:22:06,783 --> 00:22:07,783
Say something!
316
00:22:10,912 --> 00:22:12,080
Did I...
317
00:22:13,290 --> 00:22:15,709
do something wrong?
318
00:22:22,257 --> 00:22:23,342
Esaki?
319
00:22:24,593 --> 00:22:25,594
Esaki!
320
00:22:29,514 --> 00:22:30,766
Why?
321
00:22:32,809 --> 00:22:34,686
Esaki!
322
00:24:09,114 --> 00:24:14,119
Subtitle translation by: Bradley Plumb
24032
Can't find what you're looking for?
Get subtitles in any language from opensubtitles.com, and translate them here.