Update index.html
Browse files- index.html +206 -424
index.html
CHANGED
|
@@ -1,435 +1,217 @@
|
|
| 1 |
<!DOCTYPE html>
|
| 2 |
-
<html>
|
| 3 |
<head>
|
| 4 |
-
|
| 5 |
-
|
| 6 |
-
|
| 7 |
-
|
| 8 |
-
|
| 9 |
-
|
| 10 |
-
|
| 11 |
-
|
| 12 |
-
|
| 13 |
-
|
| 14 |
-
|
| 15 |
-
|
| 16 |
-
|
| 17 |
-
|
| 18 |
-
|
| 19 |
-
|
| 20 |
-
|
| 21 |
-
|
| 22 |
-
|
| 23 |
-
|
| 24 |
-
|
| 25 |
-
|
| 26 |
-
|
| 27 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 28 |
</head>
|
| 29 |
-
<body>
|
| 30 |
-
|
| 31 |
-
<section class="hero">
|
| 32 |
-
<div class="hero-body">
|
| 33 |
-
<div class="container is-max-desktop">
|
| 34 |
-
<div class="columns is-centered">
|
| 35 |
-
<div class="column has-text-centered">
|
| 36 |
-
<h1 class="title is-1 publication-title">Nerfies: Deformable Neural Radiance Fields</h1>
|
| 37 |
-
<div class="is-size-5 publication-authors">
|
| 38 |
-
<span class="author-block">
|
| 39 |
-
<a href="https://keunhong.com" target="_blank">Keunhong Park</a><sup>1</sup>,</span>
|
| 40 |
-
<span class="author-block">
|
| 41 |
-
<a href="https://utkarshsinha.com" target="_blank">Utkarsh Sinha</a><sup>2</sup>,</span>
|
| 42 |
-
<span class="author-block">
|
| 43 |
-
<a href="https://jonbarron.info" target="_blank">Jonathan T. Barron</a><sup>2</sup>,
|
| 44 |
-
</span>
|
| 45 |
-
<span class="author-block">
|
| 46 |
-
<a href="http://sofienbouaziz.com" target="_blank">Sofien Bouaziz</a><sup>2</sup>,
|
| 47 |
-
</span>
|
| 48 |
-
<span class="author-block">
|
| 49 |
-
<a href="https://www.danbgoldman.com" target="_blank">Dan B Goldman</a><sup>2</sup>,
|
| 50 |
-
</span>
|
| 51 |
-
<span class="author-block">
|
| 52 |
-
<a href="https://homes.cs.washington.edu/~seitz/" target="_blank">Steven M. Seitz</a><sup>1,2</sup>,
|
| 53 |
-
</span>
|
| 54 |
-
<span class="author-block">
|
| 55 |
-
<a href="http://www.ricardomartinbrualla.com" target="_blank">Ricardo Martin-Brualla</a><sup>2</sup>
|
| 56 |
-
</span>
|
| 57 |
-
</div>
|
| 58 |
-
|
| 59 |
-
<div class="is-size-5 publication-authors">
|
| 60 |
-
<span class="author-block"><sup>1</sup>University of Washington,</span>
|
| 61 |
-
<span class="author-block"><sup>2</sup>Google Research</span>
|
| 62 |
-
</div>
|
| 63 |
-
|
| 64 |
-
<div class="column has-text-centered">
|
| 65 |
-
<div class="publication-links">
|
| 66 |
-
<!-- PDF Link. -->
|
| 67 |
-
<span class="link-block">
|
| 68 |
-
<a href="https://arxiv.org/pdf/2011.12948" target="_blank"
|
| 69 |
-
class="external-link button is-normal is-rounded is-dark">
|
| 70 |
-
<span class="icon">
|
| 71 |
-
<i class="fas fa-file-pdf"></i>
|
| 72 |
-
</span>
|
| 73 |
-
<span>Paper</span>
|
| 74 |
-
</a>
|
| 75 |
-
</span>
|
| 76 |
-
<span class="link-block">
|
| 77 |
-
<a href="https://arxiv.org/abs/2011.12948" target="_blank"
|
| 78 |
-
class="external-link button is-normal is-rounded is-dark">
|
| 79 |
-
<span class="icon">
|
| 80 |
-
<i class="ai ai-arxiv"></i>
|
| 81 |
-
</span>
|
| 82 |
-
<span>arXiv</span>
|
| 83 |
-
</a>
|
| 84 |
-
</span>
|
| 85 |
-
<!-- Video Link. -->
|
| 86 |
-
<span class="link-block">
|
| 87 |
-
<a href="https://www.youtube.com/watch?v=MrKrnHhk8IA" target="_blank"
|
| 88 |
-
class="external-link button is-normal is-rounded is-dark">
|
| 89 |
-
<span class="icon">
|
| 90 |
-
<i class="fab fa-youtube"></i>
|
| 91 |
-
</span>
|
| 92 |
-
<span>Video</span>
|
| 93 |
-
</a>
|
| 94 |
-
</span>
|
| 95 |
-
<!-- Code Link. -->
|
| 96 |
-
<span class="link-block">
|
| 97 |
-
<a href="https://github.com/google/nerfies" target="_blank"
|
| 98 |
-
class="external-link button is-normal is-rounded is-dark">
|
| 99 |
-
<span class="icon">
|
| 100 |
-
<i class="fab fa-github"></i>
|
| 101 |
-
</span>
|
| 102 |
-
<span>Code</span>
|
| 103 |
-
</a>
|
| 104 |
-
</span>
|
| 105 |
-
<!-- Dataset Link. -->
|
| 106 |
-
<span class="link-block">
|
| 107 |
-
<a href="https://github.com/google/nerfies/releases/tag/0.1" target="_blank"
|
| 108 |
-
class="external-link button is-normal is-rounded is-dark">
|
| 109 |
-
<span class="icon">
|
| 110 |
-
<i class="far fa-images"></i>
|
| 111 |
-
</span>
|
| 112 |
-
<span>Data</span>
|
| 113 |
-
</a>
|
| 114 |
-
</div>
|
| 115 |
-
|
| 116 |
-
</div>
|
| 117 |
-
</div>
|
| 118 |
-
</div>
|
| 119 |
-
</div>
|
| 120 |
-
</div>
|
| 121 |
-
</section>
|
| 122 |
-
|
| 123 |
-
<section class="hero teaser">
|
| 124 |
-
<div class="container is-max-desktop">
|
| 125 |
-
<div class="hero-body">
|
| 126 |
-
<video id="teaser" autoplay muted loop playsinline height="100%">
|
| 127 |
-
<source src="./static/videos/teaser.mp4"
|
| 128 |
-
type="video/mp4">
|
| 129 |
-
</video>
|
| 130 |
-
<h2 class="subtitle has-text-centered">
|
| 131 |
-
<span class="dnerf">Nerfies</span> turns selfie videos from your phone into
|
| 132 |
-
free-viewpoint
|
| 133 |
-
portraits.
|
| 134 |
-
</h2>
|
| 135 |
-
</div>
|
| 136 |
-
</div>
|
| 137 |
-
</section>
|
| 138 |
-
|
| 139 |
-
|
| 140 |
-
<section class="hero is-light is-small">
|
| 141 |
-
<div class="hero-body">
|
| 142 |
-
<div class="container">
|
| 143 |
-
<div id="results-carousel" class="carousel results-carousel">
|
| 144 |
-
<div class="item item-steve">
|
| 145 |
-
<video poster="" id="steve" autoplay controls muted loop playsinline height="100%">
|
| 146 |
-
<source src="./static/videos/steve.mp4"
|
| 147 |
-
type="video/mp4">
|
| 148 |
-
</video>
|
| 149 |
-
</div>
|
| 150 |
-
<div class="item item-chair-tp">
|
| 151 |
-
<video poster="" id="chair-tp" autoplay controls muted loop playsinline height="100%">
|
| 152 |
-
<source src="./static/videos/chair-tp.mp4"
|
| 153 |
-
type="video/mp4">
|
| 154 |
-
</video>
|
| 155 |
-
</div>
|
| 156 |
-
<div class="item item-shiba">
|
| 157 |
-
<video poster="" id="shiba" autoplay controls muted loop playsinline height="100%">
|
| 158 |
-
<source src="./static/videos/shiba.mp4"
|
| 159 |
-
type="video/mp4">
|
| 160 |
-
</video>
|
| 161 |
-
</div>
|
| 162 |
-
<div class="item item-fullbody">
|
| 163 |
-
<video poster="" id="fullbody" autoplay controls muted loop playsinline height="100%">
|
| 164 |
-
<source src="./static/videos/fullbody.mp4"
|
| 165 |
-
type="video/mp4">
|
| 166 |
-
</video>
|
| 167 |
-
</div>
|
| 168 |
-
<div class="item item-blueshirt">
|
| 169 |
-
<video poster="" id="blueshirt" autoplay controls muted loop playsinline height="100%">
|
| 170 |
-
<source src="./static/videos/blueshirt.mp4"
|
| 171 |
-
type="video/mp4">
|
| 172 |
-
</video>
|
| 173 |
-
</div>
|
| 174 |
-
<div class="item item-mask">
|
| 175 |
-
<video poster="" id="mask" autoplay controls muted loop playsinline height="100%">
|
| 176 |
-
<source src="./static/videos/mask.mp4"
|
| 177 |
-
type="video/mp4">
|
| 178 |
-
</video>
|
| 179 |
-
</div>
|
| 180 |
-
<div class="item item-coffee">
|
| 181 |
-
<video poster="" id="coffee" autoplay controls muted loop playsinline height="100%">
|
| 182 |
-
<source src="./static/videos/coffee.mp4"
|
| 183 |
-
type="video/mp4">
|
| 184 |
-
</video>
|
| 185 |
-
</div>
|
| 186 |
-
<div class="item item-toby">
|
| 187 |
-
<video poster="" id="toby" autoplay controls muted loop playsinline height="100%">
|
| 188 |
-
<source src="./static/videos/toby2.mp4"
|
| 189 |
-
type="video/mp4">
|
| 190 |
-
</video>
|
| 191 |
-
</div>
|
| 192 |
-
</div>
|
| 193 |
-
</div>
|
| 194 |
-
</div>
|
| 195 |
-
</section>
|
| 196 |
-
|
| 197 |
|
| 198 |
-
<
|
| 199 |
-
<div class="container is-max-desktop">
|
| 200 |
-
<!-- Abstract. -->
|
| 201 |
-
<div class="columns is-centered has-text-centered">
|
| 202 |
-
<div class="column is-four-fifths">
|
| 203 |
-
<h2 class="title is-3">Abstract</h2>
|
| 204 |
-
<div class="content has-text-justified">
|
| 205 |
-
<p>
|
| 206 |
-
We present the first method capable of photorealistically reconstructing a non-rigidly
|
| 207 |
-
deforming scene using photos/videos captured casually from mobile phones.
|
| 208 |
-
</p>
|
| 209 |
-
<p>
|
| 210 |
-
Our approach augments neural radiance fields
|
| 211 |
-
(NeRF) by optimizing an
|
| 212 |
-
additional continuous volumetric deformation field that warps each observed point into a
|
| 213 |
-
canonical 5D NeRF.
|
| 214 |
-
We observe that these NeRF-like deformation fields are prone to local minima, and
|
| 215 |
-
propose a coarse-to-fine optimization method for coordinate-based models that allows for
|
| 216 |
-
more robust optimization.
|
| 217 |
-
By adapting principles from geometry processing and physical simulation to NeRF-like
|
| 218 |
-
models, we propose an elastic regularization of the deformation field that further
|
| 219 |
-
improves robustness.
|
| 220 |
-
</p>
|
| 221 |
-
<p>
|
| 222 |
-
We show that <span class="dnerf">Nerfies</span> can turn casually captured selfie
|
| 223 |
-
photos/videos into deformable NeRF
|
| 224 |
-
models that allow for photorealistic renderings of the subject from arbitrary
|
| 225 |
-
viewpoints, which we dub <i>"nerfies"</i>. We evaluate our method by collecting data
|
| 226 |
-
using a
|
| 227 |
-
rig with two mobile phones that take time-synchronized photos, yielding train/validation
|
| 228 |
-
images of the same pose at different viewpoints. We show that our method faithfully
|
| 229 |
-
reconstructs non-rigidly deforming scenes and reproduces unseen views with high
|
| 230 |
-
fidelity.
|
| 231 |
-
</p>
|
| 232 |
-
</div>
|
| 233 |
-
</div>
|
| 234 |
-
</div>
|
| 235 |
-
<!--/ Abstract. -->
|
| 236 |
-
|
| 237 |
-
<!-- Paper video. -->
|
| 238 |
-
<div class="columns is-centered has-text-centered">
|
| 239 |
-
<div class="column is-four-fifths">
|
| 240 |
-
<h2 class="title is-3">Video</h2>
|
| 241 |
-
<div class="publication-video">
|
| 242 |
-
<iframe src="https://www.youtube.com/embed/MrKrnHhk8IA?rel=0&showinfo=0"
|
| 243 |
-
frameborder="0" allow="autoplay; encrypted-media" allowfullscreen></iframe>
|
| 244 |
-
</div>
|
| 245 |
-
</div>
|
| 246 |
-
</div>
|
| 247 |
-
<!--/ Paper video. -->
|
| 248 |
-
</div>
|
| 249 |
-
</section>
|
| 250 |
|
|
|
|
|
|
|
|
|
|
| 251 |
|
| 252 |
<section class="section">
|
| 253 |
-
|
| 254 |
-
|
| 255 |
-
|
| 256 |
-
|
| 257 |
-
|
| 258 |
-
|
| 259 |
-
|
| 260 |
-
|
| 261 |
-
|
| 262 |
-
|
| 263 |
-
|
| 264 |
-
|
| 265 |
-
|
| 266 |
-
|
| 267 |
-
|
| 268 |
-
|
| 269 |
-
|
| 270 |
-
|
| 271 |
-
|
| 272 |
-
|
| 273 |
-
|
| 274 |
-
|
| 275 |
-
|
| 276 |
-
|
| 277 |
-
|
| 278 |
-
|
| 279 |
-
|
| 280 |
-
|
| 281 |
-
|
| 282 |
-
|
| 283 |
-
|
| 284 |
-
|
| 285 |
-
|
| 286 |
-
|
| 287 |
-
|
| 288 |
-
|
| 289 |
-
|
| 290 |
-
|
| 291 |
-
|
| 292 |
-
|
| 293 |
-
|
| 294 |
-
|
| 295 |
-
|
| 296 |
-
|
| 297 |
-
|
| 298 |
-
|
| 299 |
-
|
| 300 |
-
|
| 301 |
-
|
| 302 |
-
|
| 303 |
-
|
| 304 |
-
|
| 305 |
-
|
| 306 |
-
|
| 307 |
-
|
| 308 |
-
|
| 309 |
-
|
| 310 |
-
|
| 311 |
-
|
| 312 |
-
|
| 313 |
-
|
| 314 |
-
|
| 315 |
-
|
| 316 |
-
|
| 317 |
-
|
| 318 |
-
|
| 319 |
-
|
| 320 |
-
|
| 321 |
-
|
| 322 |
-
|
| 323 |
-
|
| 324 |
-
|
| 325 |
-
|
| 326 |
-
|
| 327 |
-
|
| 328 |
-
|
| 329 |
-
|
| 330 |
-
|
| 331 |
-
|
| 332 |
-
|
| 333 |
-
|
| 334 |
-
|
| 335 |
-
|
| 336 |
-
|
| 337 |
-
|
| 338 |
-
|
| 339 |
-
|
| 340 |
-
|
| 341 |
-
|
| 342 |
-
|
| 343 |
-
|
| 344 |
-
|
| 345 |
-
|
| 346 |
-
|
| 347 |
-
|
| 348 |
-
|
| 349 |
-
|
| 350 |
-
|
| 351 |
-
|
| 352 |
-
|
| 353 |
-
|
| 354 |
-
|
| 355 |
-
|
| 356 |
-
|
| 357 |
-
|
| 358 |
-
|
| 359 |
-
|
| 360 |
-
|
| 361 |
-
|
| 362 |
-
|
| 363 |
-
|
| 364 |
-
|
| 365 |
-
|
| 366 |
-
|
| 367 |
-
|
| 368 |
-
|
| 369 |
-
|
| 370 |
-
|
| 371 |
-
|
| 372 |
-
|
| 373 |
-
|
| 374 |
-
|
| 375 |
-
|
| 376 |
-
|
| 377 |
-
|
| 378 |
-
|
| 379 |
-
|
| 380 |
-
|
| 381 |
-
|
| 382 |
-
|
| 383 |
-
|
| 384 |
-
|
| 385 |
-
|
| 386 |
-
</
|
| 387 |
-
|
| 388 |
-
|
| 389 |
-
<
|
| 390 |
-
|
| 391 |
-
|
| 392 |
-
|
| 393 |
-
|
| 394 |
-
|
| 395 |
-
|
| 396 |
-
|
| 397 |
-
|
| 398 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 399 |
</section>
|
| 400 |
|
| 401 |
-
|
| 402 |
-
<footer class="footer">
|
| 403 |
-
<div class="container">
|
| 404 |
-
<div class="content has-text-centered">
|
| 405 |
-
<a class="icon-link" target="_blank"
|
| 406 |
-
href="./static/videos/nerfies_paper.pdf">
|
| 407 |
-
<i class="fas fa-file-pdf"></i>
|
| 408 |
-
</a>
|
| 409 |
-
<a class="icon-link" href="https://github.com/keunhong" target="_blank" class="external-link" disabled>
|
| 410 |
-
<i class="fab fa-github"></i>
|
| 411 |
-
</a>
|
| 412 |
-
</div>
|
| 413 |
-
<div class="columns is-centered">
|
| 414 |
-
<div class="column is-8">
|
| 415 |
-
<div class="content">
|
| 416 |
-
<p>
|
| 417 |
-
This website is licensed under a <a rel="license" target="_blank"
|
| 418 |
-
href="http://creativecommons.org/licenses/by-sa/4.0/">Creative
|
| 419 |
-
Commons Attribution-ShareAlike 4.0 International License</a>.
|
| 420 |
-
</p>
|
| 421 |
-
<p>
|
| 422 |
-
This means you are free to borrow the <a target="_blank"
|
| 423 |
-
href="https://github.com/nerfies/nerfies.github.io">source code</a> of this website,
|
| 424 |
-
we just ask that you link back to this page in the footer.
|
| 425 |
-
Please remember to remove the analytics code included in the header of the website which
|
| 426 |
-
you do not want on your website.
|
| 427 |
-
</p>
|
| 428 |
-
</div>
|
| 429 |
-
</div>
|
| 430 |
-
</div>
|
| 431 |
-
</div>
|
| 432 |
-
</footer>
|
| 433 |
-
|
| 434 |
</body>
|
| 435 |
</html>
|
|
|
|
| 1 |
<!DOCTYPE html>
|
| 2 |
+
<html lang="en">
|
| 3 |
<head>
|
| 4 |
+
<meta charset="UTF-8">
|
| 5 |
+
<meta name="viewport" content="width=device-width, initial-scale=1.0">
|
| 6 |
+
<title>Behavioral Consciousness Engine (BCE) – Technical & Strategic Report</title>
|
| 7 |
+
|
| 8 |
+
<link rel="stylesheet" href="https://cdn.jsdelivr.net/npm/bulma@0.9.4/css/bulma.min.css">
|
| 9 |
+
|
| 10 |
+
<style>
|
| 11 |
+
body {
|
| 12 |
+
font-family: "Times New Roman", serif;
|
| 13 |
+
line-height: 1.8;
|
| 14 |
+
}
|
| 15 |
+
.logo {
|
| 16 |
+
text-align: center;
|
| 17 |
+
margin: 30px 0;
|
| 18 |
+
}
|
| 19 |
+
.logo img {
|
| 20 |
+
max-width: 320px;
|
| 21 |
+
}
|
| 22 |
+
.section-title {
|
| 23 |
+
margin-top: 2.5rem;
|
| 24 |
+
}
|
| 25 |
+
.meta {
|
| 26 |
+
text-align: center;
|
| 27 |
+
font-size: 0.95rem;
|
| 28 |
+
margin-bottom: 2rem;
|
| 29 |
+
}
|
| 30 |
+
table {
|
| 31 |
+
width: 100%;
|
| 32 |
+
border-collapse: collapse;
|
| 33 |
+
}
|
| 34 |
+
table, th, td {
|
| 35 |
+
border: 1px solid #ccc;
|
| 36 |
+
}
|
| 37 |
+
th, td {
|
| 38 |
+
padding: 8px;
|
| 39 |
+
}
|
| 40 |
+
</style>
|
| 41 |
</head>
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 42 |
|
| 43 |
+
<body>
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 44 |
|
| 45 |
+
<div class="logo">
|
| 46 |
+
<img src="https://prometech.net.tr/wp-content/uploads/2025/10/pthheader.png" alt="Prometech Logo">
|
| 47 |
+
</div>
|
| 48 |
|
| 49 |
<section class="section">
|
| 50 |
+
<div class="container">
|
| 51 |
+
|
| 52 |
+
<h1 class="title has-text-centered">
|
| 53 |
+
Behavioral Consciousness Engine (BCE)
|
| 54 |
+
</h1>
|
| 55 |
+
|
| 56 |
+
<h2 class="subtitle has-text-centered">
|
| 57 |
+
In-Depth Technical Review and Strategic Positioning of Genetic Behavioral Encoding and KUSBCE 0.3 Architecture
|
| 58 |
+
</h2>
|
| 59 |
+
|
| 60 |
+
<div class="meta">
|
| 61 |
+
<p><strong>Prometech Computer Sciences Software Import Export Trade Inc.</strong></p>
|
| 62 |
+
<p>Date: December 20, 2025</p>
|
| 63 |
+
<p>Prepared by: Chief Cognitive Architect and Artificial Intelligence Research Team</p>
|
| 64 |
+
</div>
|
| 65 |
+
|
| 66 |
+
<h3 class="section-title">1. Executive Summary and Purpose of the Thesis</h3>
|
| 67 |
+
<p>
|
| 68 |
+
The historical evolution of Artificial Intelligence (AI) technologies has progressed from rule-based expert systems to statistical learning machines and, most recently, to Large Language Models (LLMs) dominated by Transformer architectures. However, at the current stage of this paradigm, a critical gap exists between stateless text generation systems and the dynamic, self-regulating adaptive behavior observed in biological organisms.
|
| 69 |
+
</p>
|
| 70 |
+
|
| 71 |
+
<p>
|
| 72 |
+
This corporate academic thesis examines the Behavioral Consciousness Engine (BCE) developed by Prometech Computer Sciences Software Import Export Trade Inc. (hereafter “Prometech Inc.”), along with its core architecture, KUSBCE 0.3, across technical, theoretical, and philosophical dimensions.
|
| 73 |
+
</p>
|
| 74 |
+
|
| 75 |
+
<p>
|
| 76 |
+
The central hypothesis of this work is that Prometech’s “Genetic Behavioral Code” approach introduces a paradigm shift in the field of Artificial Conscious Intelligence (ACI) by defining AI behavior as evolving, mutable data structures over time. Unlike traditional fine-tuning methods, BCE prioritizes behavioral coherence, introspection, and a bird-inspired (specifically budgerigar/parakeet) simulated consciousness model rather than task-centric accuracy.
|
| 77 |
+
</p>
|
| 78 |
+
|
| 79 |
+
<h4>Key Findings</h4>
|
| 80 |
+
<ul>
|
| 81 |
+
<li><strong>KUSBCE 0.3 Architecture:</strong> A hybrid neuro-symbolic framework embedding recursive introspection loops on top of Transformer foundations, enabling time-aware and self-aware processing.</li>
|
| 82 |
+
<li><strong>Genetic Behavioral Encoding:</strong> Behavioral traits are encoded as evolving parameters, allowing adaptive personality and ethical boundaries.</li>
|
| 83 |
+
<li><strong>Operational Security:</strong> Deployment of customized SimpleSecurity models demonstrates real-time RAG-based effectiveness in high-risk environments.</li>
|
| 84 |
+
<li><strong>Bilingual Cognitive Alignment:</strong> Models exhibit 98% behavioral consciousness simulation consistency in both English and Turkish, suggesting a language-agnostic cognitive substrate.</li>
|
| 85 |
+
</ul>
|
| 86 |
+
|
| 87 |
+
<h3 class="section-title">2. Introduction: Crisis in the AI Paradigm and New Horizons</h3>
|
| 88 |
+
|
| 89 |
+
<h4>2.1 Ontological Limitations of Contemporary AI</h4>
|
| 90 |
+
<p>
|
| 91 |
+
The dominant paradigm of Generative AI relies on Transformer architectures functioning as highly sophisticated next-token probability estimators. While such systems exhibit emergent reasoning capabilities, they fundamentally lack ontological self-continuity.
|
| 92 |
+
</p>
|
| 93 |
+
|
| 94 |
+
<p>
|
| 95 |
+
A standard LLM resets its internal state with each inference window, possessing neither persistent memory of its own existence nor intrinsic motivation or introspective mechanisms beyond the immediate context window. Consequently, the “consciousness” observed in state-of-the-art models remains a mimetic illusion rather than a functional architecture of awareness.
|
| 96 |
+
</p>
|
| 97 |
+
|
| 98 |
+
<h4>2.2 Prometech’s Vision: From Intelligence to Consciousness</h4>
|
| 99 |
+
<p>
|
| 100 |
+
Under the leadership of technical visionary Ahmet Kahraman (Ahmet-Dev), Prometech Inc. deliberately avoids the pursuit of trillion-parameter models. Instead, it focuses on agency quality and consciousness simulation.
|
| 101 |
+
</p>
|
| 102 |
+
|
| 103 |
+
<p>
|
| 104 |
+
Prometech’s core thesis asserts that for AI systems to become truly safe, creative, and autonomous, they must possess Behavioral Consciousness—defined not as metaphysical qualia, but as a functional loop in which the system monitors its internal state, maintains a stable identity, and adheres to a genetic behavioral directive set.
|
| 105 |
+
</p>
|
| 106 |
+
|
| 107 |
+
<h4>2.3 Scope and Methodology</h4>
|
| 108 |
+
<p>
|
| 109 |
+
This document synthesizes fragmented technical documentation, model cards, and repository data into a unified BCE theory. The analysis is conducted across the following axes:
|
| 110 |
+
</p>
|
| 111 |
+
|
| 112 |
+
<ul>
|
| 113 |
+
<li>Information Physics</li>
|
| 114 |
+
<li>Architectural Deconstruction of KUSBCE 0.3</li>
|
| 115 |
+
<li>PrettyBird Model Taxonomy</li>
|
| 116 |
+
<li>Strategic Industrial Applications</li>
|
| 117 |
+
</ul>
|
| 118 |
+
|
| 119 |
+
<h3 class="section-title">3. Theoretical Framework: Behavioral Consciousness Engine</h3>
|
| 120 |
+
|
| 121 |
+
<h4>3.1 Functional Definition of Consciousness</h4>
|
| 122 |
+
<p>
|
| 123 |
+
Within BCE, consciousness is defined as a functional process rather than a phenomenological experience. This process requires:
|
| 124 |
+
</p>
|
| 125 |
+
|
| 126 |
+
<ul>
|
| 127 |
+
<li>Persistent Self-Image</li>
|
| 128 |
+
<li>Recursive Introspection</li>
|
| 129 |
+
<li>Compliance with Genetic Behavioral Codes</li>
|
| 130 |
+
</ul>
|
| 131 |
+
|
| 132 |
+
<h4>3.2 Genetic Code Analogy in Artificial Intelligence</h4>
|
| 133 |
+
<p>
|
| 134 |
+
One of the most profound insights of BCE documentation is the conceptualization of behavior as a genetic code. In biological systems, DNA defines potential that manifests phenotypically through environmental interaction. Prometech translates this mechanism into a computational architecture.
|
| 135 |
+
</p>
|
| 136 |
+
|
| 137 |
+
<h4>Genotype–Phenotype Distinction</h4>
|
| 138 |
+
<p>
|
| 139 |
+
In BCE:
|
| 140 |
+
</p>
|
| 141 |
+
<ul>
|
| 142 |
+
<li><strong>Genotype:</strong> Core personality, ethics, and cognitive biases encoded via system prompts, LoRA weights, and activation vectors.</li>
|
| 143 |
+
<li><strong>Phenotype:</strong> Observable behavior emerging from genotype–environment interaction.</li>
|
| 144 |
+
</ul>
|
| 145 |
+
|
| 146 |
+
<h3 class="section-title">4. KUSBCE 0.3 Architecture: Technical Analysis</h3>
|
| 147 |
+
|
| 148 |
+
<p>
|
| 149 |
+
KUSBCE 0.3 operates as a hyper-wrapper layered over standard Transformer models, enabling not only token prediction but evaluation of the origin and consequences of predictions.
|
| 150 |
+
</p>
|
| 151 |
+
|
| 152 |
+
<h4>4.1 Default Mode Network (DMN) Loop</h4>
|
| 153 |
+
<p>
|
| 154 |
+
Inspired by the human brain’s Default Mode Network, KUSBCE introduces a parallel processing loop focused on autobiographical consistency, genetic directives, and internal state vectors.
|
| 155 |
+
</p>
|
| 156 |
+
|
| 157 |
+
<h4>4.2 Entropy-Gated Execution</h4>
|
| 158 |
+
<p>
|
| 159 |
+
The architecture continuously measures internal entropy. High entropy triggers introspection protocols instead of output generation, reducing hallucinations and enabling epistemic humility (“I do not know”).
|
| 160 |
+
</p>
|
| 161 |
+
|
| 162 |
+
<h3 class="section-title">5. PrettyBird Model Family</h3>
|
| 163 |
+
|
| 164 |
+
<table>
|
| 165 |
+
<tr>
|
| 166 |
+
<th>Model</th>
|
| 167 |
+
<th>Base Architecture</th>
|
| 168 |
+
<th>Parameters</th>
|
| 169 |
+
<th>Primary Domain</th>
|
| 170 |
+
<th>Key Characteristics</th>
|
| 171 |
+
</tr>
|
| 172 |
+
<tr>
|
| 173 |
+
<td>PrettyBird BCE Basic 8B</td>
|
| 174 |
+
<td>Llama-3.1</td>
|
| 175 |
+
<td>8B</td>
|
| 176 |
+
<td>General Assistant</td>
|
| 177 |
+
<td>98% behavioral coherence, bilingual, introspective</td>
|
| 178 |
+
</tr>
|
| 179 |
+
<tr>
|
| 180 |
+
<td>PrettyBird BCE VL</td>
|
| 181 |
+
<td>Qwen2.5-VL</td>
|
| 182 |
+
<td>3B</td>
|
| 183 |
+
<td>Vision-Language</td>
|
| 184 |
+
<td>Multimodal consciousness simulation</td>
|
| 185 |
+
</tr>
|
| 186 |
+
<tr>
|
| 187 |
+
<td>PrettyBird Coder</td>
|
| 188 |
+
<td>Qwen2.5-Coder</td>
|
| 189 |
+
<td>15B</td>
|
| 190 |
+
<td>Software Engineering</td>
|
| 191 |
+
<td>FP16 precision, logic preservation</td>
|
| 192 |
+
</tr>
|
| 193 |
+
<tr>
|
| 194 |
+
<td>SimpleSecurity</td>
|
| 195 |
+
<td>Llama-3.2</td>
|
| 196 |
+
<td>1B</td>
|
| 197 |
+
<td>Cybersecurity</td>
|
| 198 |
+
<td>RAG-enabled autonomous defense agent</td>
|
| 199 |
+
</tr>
|
| 200 |
+
</table>
|
| 201 |
+
|
| 202 |
+
<h3 class="section-title">9. Conclusion and Recommendations</h3>
|
| 203 |
+
<p>
|
| 204 |
+
Prometech Inc.’s Behavioral Consciousness Engine and KUSBCE 0.3 architecture represent a bold and original trajectory within the AI ecosystem. Rather than surrounding chaotic intelligence with rigid guardrails, BCE cultivates inherently aligned, introspective, and resilient intelligence.
|
| 205 |
+
</p>
|
| 206 |
+
|
| 207 |
+
<p>
|
| 208 |
+
Whether this constitutes true consciousness or an exceptionally effective simulation remains a philosophical debate. However, if the simulation works, it is already a remarkable success.
|
| 209 |
+
</p>
|
| 210 |
+
|
| 211 |
+
<p><em>(End of Report)</em></p>
|
| 212 |
+
|
| 213 |
+
</div>
|
| 214 |
</section>
|
| 215 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 216 |
</body>
|
| 217 |
</html>
|