Killing Windows Night Light

O.M.F.G I finally found out why my colour grading has been so off. Windows 10 “kindly” enabled it’s Night Light mode and made everything more Red in the evening.

Night Light mode is the same as F.Lux or Twilight Mode it puts a Red colour over the top of your screen to reduce the amount of blue light, this is meant to help you go to sleep easier.

However, when doing any colour grading work it completely throws off your attempts. You need to disable it in order to do any Photoshopping, Video editing or anything to do with colour grading.

You can go to [Settings] -> [Display]
Then check that [Night light] is switched to off.

Then go to [Night light settings]

In the Night light settings ensure that the Schedule is set to Off and that the Colour temperature at night is all the way to the right and thus is white.

Unfortunately it took me way too long to realise what was going on and why. When I colour corrected a clip and it looked white to me, the Davinci Resolve colour scopes looked like they were out. I’d notice what looked like a bit of Red on the monitor, but would just tilt it until it looked fine. Which seemed fine as I had an X-Rite i1 Display Pro colour calibrator and was using a 4K monitor.

Note that there’s also a few other apps which can cause a colour tint. On my Asus Republic of Gamers laptop the Armoury Crate app includes a Featured app called GameVisual which also likes to do some colour temperature changes of its own.

Hopefully this helps others with some similar issues. Let me know of any other apps which causes problems.

Things I want to teach my children

Here’s a collection of things I’d love to teach my kids:

One of the most important things is how to be happy and successful. The secret to happiness being more than just having low expectations and being happily surprised.
Having a meaningful life where you are working towards a unified purpose is important.

Another thing is not to take on too many things. I’ve found it hard to say no, because I can see that there’s not enough skilled people in the world trying to help and that I can see the potential in many of the projects I come across.
But by making a list of 25+ things you want to do and focusing on the top 5 you can hopefully keep your focus.

Note that to know your purpose and thus how to prioritise what you want to do in life you should read Stephen R. Covey’s book, The 7 Habits of Highly Effective People which will also teach you to do things like balance production vs production capacity as well as taking different views of your life and do decent planning

vagrant plugin update

Vagrant is a program which makes it easy to start new Virtual Machines.
I’ve got a Windows machine (for the games and video editing software). But usually code websites which are run on Linux servers.

I usually have 1 or 2 VM’s running on my laptop.

After getting messages from Vagrant every time I started up a VM that a new update was available, I decided to install the latest version (v2.2.5)

It then stopped working and when running the usual vagrant up I got the following:

Vagrant failed to initialize at a very early stage:
The plugins failed to initialize correctly. This may be due to manual
modifications made within the Vagrant home directory. Vagrant can
attempt to automatically correct this issue by running:
vagrant plugin repair
If Vagrant was recently updated, this error may be due to incompatible
versions of dependencies. To fix this problem please remove and re-install
all plugins. Vagrant can attempt to do this automatically by running:
vagrant plugin expunge –reinstall
Or you may want to try updating the installed plugins to their latest
vagrant plugin update
Error message given during initialization: Unable to resolve dependency: user requested ‘vagrant-hostmanager (= 1.8.9)’

Running a vagrant plugin repair showed a new error.

Unable to resolve dependency: user requested ‘vagrant-vbguest (= 0.18.0)

Running the vagrant plugin expunge –reinstall didn’t help.

The vbguest is a reference to the VirtualBox manager I use and likely the guest plugins which allow for better 2 way communication between my host machine and the guest VMs.

There was no reference to the plugin in my Vagrant file, nor in the vagrant folder. There wasn’t any good Google results (hence why I’m writing this post).

After some playing around I found the command which fixed it:

vagrant plugin update

Running a vagrant plugin update caused it to update to v0.19.0 of the plugin and then everything worked happily.

Hopefully if others have the same issue they can quickly try a vagrant plugin update and see if that fixes their issue.

Yii2 Swiftmailer 0 Auth exception

If you get the error Message: Failed to authenticate on SMTP server with username “****” using 0 possible authenticators

Then try to remove the username and password from the configuration file.


When using the Swiftmailer, a common PHP emailer
This example specifically talks about the Yii2 configuration file, but likely applies to other frameworks.

Here’s an example of the offending config

config/web.php (or console.php or a common.php file if you merge the two).

'components' => [
  'mailer' => [
    'class' => 'yii\swiftmailer\Mailer',
    'transport' => [
        'class' => 'Swift_SmtpTransport',
        'plugins' => [
            ['class' => 'Openbuildings\Swiftmailer\CssInlinerPlugin']
        "username" => "smtp-auth-user",
        "password" => "*****",
        "host" => 'exchange.local',
        "port" => 25,

The exception seen was

Message: Failed to authenticate on SMTP server with username “….” using 0 possible authenticators

This exception caused major headache.

After investigation it turned out that removing the username and password from the transport caused it to work.

It seems that the server we were on was in a corporate environment and SMTP authentication was disabled but Swiftmailer was trying to authenticate and failed.

Bonus – Enabling SMTP Logging

'components' => [
'mailer' => [
'class' => 'yii\swiftmailer\Mailer',
'enableSwiftMailerLogging' => true,
'transport' => [
'class' => 'Swift_SmtpTransport',
"host" => 'localhost',
"port" => 25,
'log' => [
'traceLevel' => YII_DEBUG ? 3 : 0,
'targets' => [
'class' => 'yii\log\FileTarget',
'levels' => ['error', 'warning'],
// Log the emails
'class' => 'yii\log\FileTarget',
'categories' => ['yii\swiftmailer\Logger::add'],
'logFile' => '@app/runtime/logs/email.log',


With the above config you should now see detailed logs in the runtime/logs/email.log file.

Sabby Love

I’m a bit of a night owl but she takes it to a whole new level, sleeping most of the day.
So when she started staying over we got very little sleep. I was exhausted. But we are both in a good rhythm now.
I love the way we’ll seek each other and curl up next to each other. She’ll fall asleep in my arms. Other times she’ll caress my feet. Sometimes she’ll bite and scratch a bit when she’s feeling that way inclined. It’s not my thing, but to each their own.
I met her some time ago but we met again through the same mutual friend who looks after her Mum and we’ve been together for some months now. She’s black, which is new for me, although she has some white hairs that really stand out and I occasionally pluck out.
I love her.
Her eating habits leave a lot to be desired. Like many Vietnamese, they don’t put their litter in the bin but on the floor so I often have to sweep up afterwards.
We’ve already been through a lot. When she first arrived she was a scaredy cat, especially afraid of the rain and thunder. The roof here does make the rain extra loud 🔊  but now she is fine and can sleep through it.
It’s hard to have any privacy with her around. She’ll often come into the bathroom whilst I’m sitting on the toilet. Yet she keeps away when I shower.

We’ve watched movies together and fought off the flies, moths and bugs that often try to attack in night. She loves it when I sweep up. Did didn’t always like me on the computer too long, but now she enjoys it because she’s learnt how to be with me.

I love watching her play and enjoy life. She’s so cute.
Although she’s also invasive. She’ll check everything and go through my stuff given half a chance and sometimes destroys things. But she’s curious, not malicious.
But. I recently learnt that she’s not a she.
It turns out that Sabby, my cat, is a boy. Both Mrs Loan and myself misinterpreted Sabby’s gender. Apparently that’s pretty easy to do when the cat is young.
Sabby is a mostly black Bombay Cat whom I love dearly.
I was going to post this some time ago but wanted to make a nice collage of images. I ran out of time then, but in 15 minutes time Sabby is going away, back to his home.
See, I’ve been living in Vietnam for nearly 7 months now, but it’s time to return to Australia and then, well I’m not sure where I’m going next but I know it’s to be with my girlfriend Jen.

Using jq to update the contents of certain JSON fields

OK, I’ll be brief.


I created a set of API docs with Apiary using Markdown format.

We needed to change over to Postman, so I used Apimatic for the conversion. Which was 99% great, except for the item descriptions it only did a single line break, not two line breaks. As Postman is reading the description as Markdown a single line break doesn’t actually create a new line.


So, I needed to replace the string \n with \n\n but the key is I only needed to do it on the description field.

Ohh and I needed to add an x-api-key to use the mock server. Even Postman’s own authorisation system didn’t seem to easily support this.

Using the incredibly useful comment by NathanNorman on this GitHub Postman issue I had a glimpse of what I could do.


So to add in the x-api-key into the  Postman headers, on my Linux VM I ran the following on the terminal:

jq ‘walk(if (type == “object” and has(“header”)) then .header |= (. + [{“key”:“x-api-key”, “value”:“{{apiKey}}”}] | unique) else . end )’ postman_api.json > postman_api_apiHeader.json


I then checked some resources, learnt about the |= update operator and gsub for replacement.

So to replace \n with \n\n in just the description fields I ended up with:

cat postman_api_apiHeader.json | jq ‘walk( if (type == “object” and has (“description”) ) then .description |= gsub( “\\n”; “\n\n”) else . end )’ > postman_api_apiHeader_description.json


If you want to see a list of the updated description fields to make sure it worked you can pipe the results to jq again.

cat postman_api_apiHeader.json | jq ‘walk( if (type == “object” and has (“description”) ) then .description |= gsub( “\\n”; “\n\n”) else . end )’ | jq ‘..|.description?’

Hopefully that helps others, or myself in the future.


Note that I downloaded the latest version of jq in order to run this. The debian distros are only using version 1.5 but I needed v1.6 for the walk function, but it’s a pretty easy download.

Some resources: Official jq site – Official docs description of the Walk function in jq. – Some jq recipes – The Github issue that got me down this path – The very powerful API blueprint online conversion system. Allowing me to upload a Markdown style Apiary file and download a Postman and also Swagger .json files.



Fake Taxi’s in Vietnam

I’m writing this because I paid 10x the price I should have for a fairly short taxi ride.

This is based on my one bad experience after months of living in Vietnam.

Firstly. If you are going to get in a taxi you should have the Grab app. It’s their equivalent to Uber. Even if you don’t want to, or for some reason can’t order a Grab, look at the suggested price. The taxi prices are usually within 30% of that price, actually they are usually almost identical.
It might be a little different for longer trips, but usually you have to be a hotel to negotiate good prices to the airport that are better than the Grab or normal taxi rates.

So Grab gives you a good reference point price.

The Meter

The second thing to look at is the meter.

The fake Taxi meter looked like this

A fake Taxi Meter
A fake Taxi Meter

See it’s just got a single display field. The price.

The real Taxi meters look like this

A proper taxi meter.
A proper taxi meter.

The top left most number is the price

Another photo of a proper taxi meter (with flash so you can see the front face better)
Another photo of a proper taxi meter (with flash so you can see the front face better)

See the 4 different fields. They don’t just show a price. There’s Fare, Time, Distance and Unit Price. There’s also 5 buttons on it.

For the fake taxi trip I traveled for less than 5 minutes, just over 2km and paid $420,000 VND (Vietnamese Dong) or about AUD$25. It should have cost 40k not 450k.

For the legit taxi ride as a reference when I had travelled for 10km it showed 121k VND.

The fake taxi itself looked fairly legit from the brief walk up to it I did, it had been painted green and had a thing on top. He was positioned as if he just dropped a passenger off.

If you think you are being scammed then get the taxi to take you to your hotel or even just a Circle K / 7 eleven equivalent and say you don’t have the money but you’ll get it. Go and talk to the people at the hotel or store and let them know the price the Taxi is asking. See if they think it’s reasonable. If not, they can usually help out.

If language is an issue, use the Google Translate App. It’ll do conversation mode if it has Internet access and as simcards are cheap and you can buy them from the airport, this shouldn’t be an issue. Otherwise be prepared and have the Vietnamese dictionary downloaded for offline use in the app and a Vietnamese keyboard option on your phone. With TouchPal I just swipe on my space bar to change the language.

My story

I got done by the fake taxi because I’d used up my mobile phone data on the train. I have been living and teaching English in a rural village in Nghi Loc. It’s a 7 hour train ride from the nearby train station in Vinh city to Hanoi. I was going there to meet my girlfriend who was flying in from overseas.

I’ve done the trip to and from Hanoi a couple of times, but that was some months ago and I have managed to spend very little money in the meantime. My food and accomodation being covered by the BlueSky English Language Center. As such I wasn’t used to the prices.

I walked out of the train station and was asked by someone of I wanted a Taxi, I fobbed him off and instead headed to the toilet with my heavy bags.

It was around 8pm by now and after buying some food I realised I couldn’t order a Grab because of the lack of data. I should’ve got a recharge at the Circle K I was just at, but didn’t know that was an option. There’s no free Wi-Fi on the train or at the train station, not that I could find. Instead, carrying heavy camera gear and clothes for filming a wedding, I went looking for a taxi. It was a short trip to the Hanoi Lotus Hostel where I was staying. I figured that the difference between a Taxi and Grab would be barely anything.

I saw someone standing by a legit looking Taxi. The sun had set half an hour ago so it was dark, but it looked like a painted, proper Taxi with the 🚕 augment to the roof.
I got in, showing the business card of the place I was staying and he drove me to the intersection about 10m away and pointed in the wrong direction as if he thought I’d asked to be dropped off at a different location and he was stopped between the two. No big deal I thought.

The meter said 420 and I assumed the dot wasn’t displaying and it meant 42.0 or 42,000 Vietnamese Dong. About the right price I’d expected. I gave him $50k VND expecting change and he said no, 420k.

I was tired and confused. When did Hanoi get so expensive, that was more than the 7 hour train ride I’d just done.

But haven’t I paid a lot for a taxi trip before? Yeah. OK. Whatever, I’m so close, I just want to get to the hostel.

I gave him a 500k note. He gave me 20k change. Woo I said, where’s the rest. I wasn’t THAT out of it. He nearly stiffed me on the change as well, which is more than the price of a legit trip. In hindsight the slight nervous reaction I had here showed he was probably worried the gig was up. Unfortunately I got out and it wasn’t until about a minute later realised I’d just paid the most expensive taxi ride I’ve done in Vietnam, which was also my shortest trip.

So at the critical moment, when going to pay the System 1, fast processing part of my brain remembered that I’ve paid a lot for a taxi ride before, It remembered that it’s some multiple of the train ride.
But, what it didn’t remember was the specifics. The taxi ride was from Vinh city to the rural area in Nghi Loc where I teach and live. That takes nearly 30 mins and costs 250k VND. Expensive yes, but the train ticket still cost about twice that. Or the taxi is about 0.5x the cost of a VIP sleeper trip.
So my brain got the orders of magnitude out of whack and I didn’t think about it properly. I also didn’t expect to get gauged so badly.

Hopefully after reading this you now have a better chance of not being ripped off.

Some other notes

  • It only cost me 25k VND to get from the Old Quarters area to the Hanoi Railway station.
    Also, the railway station isn’t the easiest to find in the Grab search. Try “Ga Ha Noi” as terms like Hanoi Railway station or Hanoi train station don’t really work (not when I tried on the 31st of Oct 2018).

    Example Grab ride to Ga Ha Noi — The Railway station
    Example Grab ride to Ga Ha Noi — The Railway station
  • Expect it to cost 250k VND from the Old Quarters area to the Hanoi Airport. It’s a long trip with sections that can go at 80km/hr. Those are faster by car, not bike. Unless traffic is really bad getting out of the rabbit warren that is Hanoi. But trust me, don’t do the whole trip by bike. You won’t enjoy it.

    When searching for the International Terminal at Hanoi, scroll down past the Domestic ones.
    When searching for the International Terminal at Hanoi, scroll down past the Domestic ones.
  • The Hanoi Airport International terminal is listed below the domestic terminals in Grab. If you find yourself at the Domestic Terminal T1 (A or B) and need to get to T2, it’s about 2km away to the west section of the airport, not easily shown on Google Maps unless you go into satellite view. The taxi’s won’t want to take you such a short distance. From the T1 building which is Domestic Departures you head West and see the Domestic Arrivals (E), head down and go out the airport and around to the International Terminal. It took me over half an hour to work that out in a frantic hurry.

    Hanoi Airport Layout
    Hanoi Airport Layout

Stock Footage Intro

What is it?
Stock Footage is often termed generic or B roll footage. It’s used to fill in shots. Maybe a movie or TV show needs a cinematic fly over of New York city, or a corporate video needs a shot that means “Enthusiasm”. Or just footage of a dump truck, rice field, person at the top of a mountain, you get the idea.
A lot of the Zeitgeist films used stock footage. A lot of the Jay Shetty videos is stock.
Where can you buy/sell it from
There’s a lot of different stock websites.
There’s probably other drone specific websites as well that have sprung up recently.
General Information
Stock footage is almost always 5s to a max of 1 minute long, it’s without audio and is best in 4k resolution although 1080p is also accepted.
There’s some great options, from people, to nature, also things like slow-mo, timelapse, drone footage, 360° and more.
The main issues to consider when filming stock footage is the requirements for no visible branding, shots that have meaning and the need for Model and property release forms.
No Branding
If you film a person up close and their clothing has brand logos on it, you can’t sell that footage commercially. It’s considered distracting, but also there’s trademark and copyright issues. The same is the case for filming a generic shot of a shopping mall, there’s brand logos all over the place.
Even focusing too close on a single car can cause problems and the footage will be rejected.
I think this is generally a lot less of a problem with drone footage.
If you fail this you’ll likely be rejected due to visible trademark.
Model Release Forms versus Editorial Content
When filming, if there’s a person who could identify themselves from seeing the footage then that footage needs a model release form.

If you are filming in certain locations then they’ll need a property release form. I got one filled out for a theme park I filmed at. If you tried to film a cultural heritage site, or somewhere you need to buy a ticket to enter then that’s a great example of where you’ll need a property release form.

I personally use the Easy Release mobile app to help me with getting model and property release forms.
There is however a 2nd type of video. Editorial content. This isn’t for movies and the like, but instead for news organisations and can also be footage you have that advertises a specific brand, e.g Coke or Grab or Nike. Those companies or even their competitors can buy the clip if they happen to like it.
Setting the submission to editorial content is also often done for things like big groups of people, etc..
Stock Footage (and stock photos) is a long game. To make enough money to quit your day job you usually need thousands of clips and of content that people want to buy. Expect $5-30 for the sale of a decent video clip, depending on the website it’s sold from. Most clips won’t sell. There’s some stats about people getting an average of like 20c to $1/month per clip. Of course it looks more like a logarithm I’m guessing. The best 20% of clips are likely worth 80% of your income.
So there’s a stock footage website where you can’t actually buy any footage from. It’s called BlackBox and it is a syndicated submission service. You upload your footage to BlackBox, fill in the various info and then they’ll submit the content on to the other main stock footage websites.
There’s more to it than that though. Because they are already dealing with the money transfers from various sites, they can also enable things like revenue splitting.
This means that I can assign say 20 or 30% of the revenue for a clip to a friend for helping curate the content. Dealing with the release forms, adding the keywords, title and other information to the videos. This is something that’s fairly easy for them to do. I’ve already uploaded the videos, they just need a laptop and can watch the video and add the info. But I’ve got a backlog of over 100 clips I’ve not published because I haven’t done this, so having someone else do it means actually getting them out there.
You could also do a revenue sharing arrangement with say a model in the video. If you filmed a dancer or a stunt driver then you could give them 50% revenue. They’ll need an account with BlackBox though.
More resources to read The general Blackbox FAQ’s – A video by Chris Hau which is basically an advert for BlackBox. – A beginners guide to selling stock. Contains another advert for Blackbox
—- – Info about how one guy managed to make a resonable amount of money selling stock. Now he runs his own stock website. – A much more realistic version of stock footage sales. He earned $6,675 in 13 months. I love his aerial shot of the Kayakers going down the rapids. Great drone footage! Netted him about $700 in sales.
The 9 tips are:
1. Keep It Quality
2. Pick a Niche
3. Include Actors
4. Exclude Branding
5. Submit Aggressively
6. Work Out Who Pays the Big Bucks
7. Flaunt Your Work
8. Get Your Analytics On
9. Dream Big — But Not Like, Too Big – You can download the model release forms from Shutterstock here. You’ll need a model release form for selling commercial video where people are identifiable in it. E.g where there’s 1 or 2 people walking along and you film them.
I would like to point out that as of the time of writing I’ve only had about 5 videos and about 20 photos actually published to a single stock website and haven’t made any sales in the month they’ve been up there. I have a lot more footage to add, but work full time programming websites, so don’t have the time to deal with keywords and the other stuff, hence why Blackbox is attractive to me right now.

SciFi concepts – nanoBlood, the positioning problem and remote sensing

Note: These are Michael Kubler’s personal notes regarding technology likely to exist in the future. It’s written as a reference guide for The Book of New Eden SciFi novel set in a post-scarcity (Abundance Centered) society. Eli is the main character in the novel.

nanoBlood is considered a form of bodyMod but is important enough to get it’s own category. Also known as Nanobot blood it is a generic term for a few different things.

All versions come with automated spider, snake and other venom antidotes. Also helps power most of the other bodyMods and a form of nanobots in the blood is how the neuroMod slowly gets assembled in the brain over the months.
v1 : Probes which provide vast amount of info on your body. E.g 2 year warning of a heart attack, instant detection and treatment of cancer, radiation therapy (which is why Eli needs it), etc.
v2 : Red blood replacement which provides far better oxygenation ability and CO2 absorption so you can hold your breath for nearly an hour or can exercise much better. Also has amazing toxin scrubbing and processing.
v3 : The immune system is augmented. Nanobots acting as advanced White cells, T cells etc. Allows for not just a super immune system, but wirelessly transferred anti-viruses. Very quick detection of new infections, diseases and viruses and the ability to transmit an internal scan to be remotely processed and and anti-virus quickly developed and downloaded. Some neural implants can do a basic AV processing but it takes longer and takes up almost all of the processing power. Note: The nanobots are created in nanobot factories, usually embedded in the bone marrow and a couple of other points around the body, nanites (self-replicating nanobots) are NOT used due to their possible ability to go rouge.
V4 : There’s usually more than 30% of the blood as nanoblood. It also has DNA scanning facilities which, with the help of a machine that’s a bit like an MRI, allows all the cells in your body to have their DNA read.


Firstly, there’s the nanobot sensors. These augment normal human senses. There’s some which are for touch (pressure), temperature, velocity of travel, through to searching for chemical markers, DNA readers, and ones that search for viruses, bacteria and the like.

It started with people needing replacement blood and plasma due to health reasons and artificial versions being created to fill this need.
But now there’s also nano replacement blood cells. In some cases, like red blood cells they are capable of orders of magnitude better increases in oxygen carrying, although don’t always have the full range of other functions, such as certain types of waste removal. Things like CO2 and lactic acid are also considerably amped.
There’s the immune system nanoblood defence where you can wirelessly download a new biological anti-virus set from the Internet and within minutes of someone on another planet getting a new form of cold virus, you can be actively able to fight it off.

There’s a few systems at play here.

You’ve got the honeypot traps. Cells designed to look enticing to potential bacteria and viruses, but are specially crafted so their exact makeup is known and they are heavily monitored. If their structure is altered then the foreign invader is very easy to identify and then analyse. Often cloud computing is used to convert the analysis into a new anti-viral signature. Some of this analysis includes specialised simulations ensuring that the identifying markers detected, if searched for and attacked, won’t also destroy human tissue.
It’s a major concern that you’ll download an anti-virus update that’s (intentionally or not) malformed and causes the nanobot immune system to start attacking and liquefying your own body. A form of nano-ebola, except not contagious. When a new AV is applied only small amounts are released into a test environment to ensure the anti-virus isn’t dangerous to you.

It should be noted. There’s no use of nanites. Self-replicating nanobots are heavily controlled and very carefully researched, but aren’t allowed for general use due to their potential power of exponential destruction.
Instead there’s nanobot factories, usually installed in your bones which create the nanoblood. You can of course get injected with booster shots as well, as happens to Eli when trying to deal with the radiation exposure.

Other concerns include getting through the blood brain barrier, especially important during the development of a neuroMod.
This can be done through either a border controlled tunnel, which are specially reinforced nanobot tunnels which have a basic verification check at least at one end allowing valid nanobots to go through.
Some nanobots are big and powerful enough to push their way through the barrier as they want, usually with it closing up behind them.

Checkout the previous post about neuroMod levels for more info.

One of the main functions of nanoblood is for energy delivery to the neuroMod and various bodyMods. Want a light in your finger? You need to give it power. Want to send wireless communications to a drone or satellite? You need power for that too.
NanoBlood provides the needed energy. Sometimes just by transferring ATP around, but other times using more advanced and more nanobot optimised energy solutions.
Of course, there’s only so much energy the human body can generate and store in reserve. Most people have plenty of reserves of energy for burst tasks. But there’s of course options for consumption or injection of purified energy solutions. Or you can get a witricity (wireless electricity) recharging system bodyMod.

Dealing with toxins and unwanted chemicals. Improved clotting and dramatically improved healing. These are other things you can do with the right nanoBlood setup.

NanoBlood Sensors of interest:
CO2 – A nanoBot sensor network can be more accurate than the existing human body. Letting you see CO2 levels throughout your body.
O2 – Humans are susceptible to Oxygen deprivation because we only sense amounts of CO2 in the blood.
CO or Carbon Monoxide – Because the main problem is the binding of CO to the hemoglobin this isn’t an issue in artificial red blood cells
Lactic Acid – Created as a byproduct of using up energy, especially during exercise.
ATP – The energy powerhouse of the body.
DNA – Being able to read the DNA of your cells. This is a very tricky endeavour and produces vast amounts of data. A few GB per cell and trillions of cells. This is usually only done occasionally for active geneMod users or in special cases, like radiation poisoning or a geneMod gone wrong.
Pressure – This is often used as a way of helping identify the nearby other nano sensors. Sometimes used to augment the normal sensation of touch, but usually in places like the brain that we don’t normally feel it.
Temperature – This is often used as a way of helping identify the nearby other nano sensors. But of course knowing your actual body’s temperature compared to the perceived temperature can help a lot.
Hormones – Detecting the levels of your various hormones.
NeuroTransmitters – Especially useful in the brain of course. A particularly important issue is knowing when there’s an excess of used up neuroTransmitter chemicals and the brain needs to flush them out, aka, go to sleep. A process that can be vastly sped up.

There’s a whole slew of chemicals to track, both good and bad, plus specialised nanobots, like those searching for cancer cells. Also ones designed specifically for your skin.

Nanoblood sensor network – The positioning issue

So most nanoBlood sensors are very basic. They have the sensor package, be it chemical, pressure, touch or something more specialised, plus a transmitter. The use of millimetre ranged radio signals saves from filling up the bloodstream with physical chemical sensors like how the body generally works. It also allows much faster signals.
The sensors have very little power output and might only sample once a second or so, but they only need the signal to travel half a centimetre at most to the nearest relay station.
NanoBot relay stations are distributed around the body. You might have half a million sensors in your index finger and a few thousand relay bots there as well. They are bigger and don’t have sensors, just transmitters and receivers. They can also specifically relay messages to the sensors. The relays usually just forward data to their nearest neighbours until the data gets to a nearby accumulator. The accumulators are approximately the size of a pea and exist in your major bones. These are usually directly wired to each other and up to the main processors. The accumulators might receive a few billion points of data a minute and can store more than 4 hours of it.

The main processors are about the size of your finger and usually installed in your collar bone. The left and right processors work together in a redundant way, allowing for one to fail and the data to still be available. These main processors are where the powerhouse of work happens, from nanoBot signal processing to neuroMod task offloading.
Whilst they have basic 10m range wireless transceivers they are also used a lot in liaising with the shoulder mounted long range Maser system (microwave laser). The directed maser transmissions are how you can communicate with a drone or satellite flying many km away. The drones give off regular pulsed beacons indicating their location and your body fires directed radio waves at the location allowing far lower powered transmissions than would normally be needed over such ranges. Also increases privacy as it’s harder to snoop.

The main issue is that the nanobot sensors are dumb. They don’t know where they are. Their signals are usually very basic. The sensor reading (e.g current temperature, or a density of chemical reading), the sensor type, a unique Id and a counter. The counter automatically increments on each transmission.

Alone, this is hard to use. But the relays can specifically request readings from nearby neighbours and can map which other sensors are nearby. So you know physical proximity to the other sensors. The relay sensors have a timing counter with better than millisecond prevision but might not know the exact date and time, just an incrementing timer. This goes with the data packets. The accumulators do have actual dateTime information. So between all of that, you can know when a reading was taken.

Now you have a massive swath of uniqueIds and sensor readings and you now need to create a basic map.’

You need to work out where those sensors are and there’s a variety of techniques including:

Ping and traceroute tests. Similar to Internet servers, by asking the sensors, relays and accumulators to reply as soon as possible you can get a judge of distance based on how long it took to reply. You can also do trace routes, and work out how far away a sensor is based on the number of hops. Did the signal go through 30 or 5,000 relays and which ones?

Another method is to make known changes and look for which sensors reflect those changes. Warm up your hand, touch your face, sit on the ground, lay on the bed. Jump up and down. Drink water. All of these will light up different sensors. A big issue is that if you don’t have a high enough level neuroMod the visual and other neocortical information which gives great resolution to many of the senses isn’t available.

A 3rd option is external scanning.

The basic version of this is simply a video camera. This is used instead of your eyes for people who don’t have a high enough level integrated neuroMod.
But there’s actual energising scanners which are often integrated into hospital beds and in MRI looking machines. They work by externally triggering the relays or sensors in specific locations and mapping the responses. If it’s just a basic 2D scanner, like laying on the bed, then both visual information and things like rolling over and normal human motion can help increase the resolution of mapping. Having a scanner on the sides, giving two axis of transmission beams gives even greater fidelity, especially for your insides like sensors in your liver which are harder to know where they are inside of your body.

The encapsulating tube readers, can also enable a full body DNA scan if the right type of nanoBots are available. These are usually energised remotely and need to enter the cells. They can’t live between cells as many other sensors can.

Sometimes you just need a basic arm that goes part way over the body, like an arm chair rest. Or simply having a specialised doorframe is enough.

Obviously some nanosensors will be stuck between or inside of cells and are in somewhat static location, like by a bone, muscle or tendon. Others are in the blood stream and are moving targets. But the moving ones can be placed near known others so can be tracked.
There’s other issues, like the sensors break down and are replaced regularly by the nanobot factories. At a slow enough rate this isn’t a problem. You have a bunch of sensors in a known area and there’s some changes over time, but the positions are generally known. However the high metabolism mode that Eli enters causes a massive increase in nanobot turnover and he needs regular injections of new nanoBot sets, causing some of the mapping to become inaccurate, hence he has a basic external scanner built into his bed which works in with the video cameras in the room.


Remote Sensing / out of body experiences

Because the nanosensors can work wirelessly out of the body, you could feel information from the drops of blood nearby.

But it goes beyond that. The mapping of sensors and sensations can be done for objects out of your body. Say a door.
The door to my room feels the air conditioning on one side and the hot Vietnam heat on the other. It feels the wind, the cat walking past or occasionally trying to scratch it. The hinges know when they are dry and need oiling again. The handle knows when it’s being used and because of the force variations likely by who. The frame and door know when they are closed or open, but also if the house has shifted and the door doesn’t quite close properly anymore. But all this sensory information could be provided to you, once you’ve got a level 7 neuroMod. You could feel the door as if it is an extension of you.
You could then feel the sensations of a tree outside. Here the sensor dust network comes into play with GPS microdots could be used to calibrate the position of nearby smart dust, plus video cameras that track occasionally IR flashing dust sensors can provide high fidelity positioning.
So you could feel the warmth of the sun on the tree. The wind in the leaves. But also the ants crawling on it’s bark. The moisture in the air and the sap leaking from the wound when a bear swiped it during a fight. The tree obviously won :p haha

You could have sensors inside the tree with tell you about the root system and soil nutrients it’s uptaking. The moisture being raised through the trunk. The CO2 it’s pulling out of the atmosphere and using as a building block for creating more plant matter. How open the stomata are and how well it’s breathing.

That’s a single door or a single tree.
But you could also abstract up and ‘feel’ a whole house or even a whole forest. You would feel different information. The forest would include not just the trees but the deer and birds and ants and bugs and decomposing nature. The ecosystem.
The house would include bedrooms and toilets and electricity, Internet, power and water. With a whole neighbourhood being equivalent to the forest.

But how large can you go? Can you abstract to a whole country? To all the oceans? To a whole planet? To a collection of planets?

This is different to experiencing the life stream of a friend, be they human or animal. Those are based on the brains sensory perceptions and are neuroMod enabled streams which include conscious processing of the stimuli.

Whilst you could also have life streams of AI’s which are based on their own processing, and experiences (conscious or otherwise). Those aren’t the same as remote sensing which is about creating a new sensory system and new interpretation system. You’d need algorithms and AI to help with the mapping and data processing and making sense of the data. Turning it into sensations that we can relate to, or helping us develop new sensations we could never have imagined.

Mods and Apps – Science Fiction Concepts for SciFi writers

Science Fiction Concepts for SciFi writers.
Set 30-50yrs in the future.

Mods and Apps

* bioMods are for specific biological enhancements. These are usually a little bit more advanced than cosmetic surgery in 2017. The standard is the spinalTap mod, an enhanced spine and skeletal upgrade (usually including knees) that dramatically reduces skeletal issues. No more easily popped discs or dislocated knees or shoulders. Basically humans haven’t finished evolving to deal with walking upright and this helps complete that. It usually has a gMod (genetic engineering) component.

* geneMods are genetic engineering changes. Usually an injected retrovirus that rewrites your DNA. Think of it like CRISPR but working on all the cells of your body. Examples of this are the ability to change your skin pigmentation between normal shades over the course of a few days. e.g From 0 – Albino white to 5 – very black. Some changes are easy, like re-enabling a number of regrowth options already in our DNA. So you can cut your arm off and over the course of a few months it’ll grow back. Want to have your skin be pigmented to look like a purple dragon 🐉 and you are beyond normal gMods. But then, most people would just have an eInk tatto for that instead of changing those specific skin cells pigmentation.

* bodyMods are mainly physical implants of technology. Things like an LED light in your fingertip to having geiger counters or EM detectors built into your body as new senses. Want to read someone’s DNA by shaking their hand, or remove your stomach and just have print cartridge like nurtient containers that you replace? Sure. You can have the standard dental armour upgrade so you only have to brush your teeth once a year, or eInk skin that turns your body into basically a digital screen. You could become a chest breather, replacing your lungs with two holes just above your collarbone. Air goes in one hole and out the other, thus you no longer have a normal breathing motion. There’s also the usual assortment of faster legs and arms. Maybe your full cybernetic arm could have a nanofactory in it and you could leave a trail of smart dust.

* nanoBlood – Nanobot blood.
All versions come with automated spider, snake and other venom antidotes. Also helps power most of the other bodyMods and a form of nanobots in the blood is how the neuroMod slowly gets assembled in the brain over the months.
v1 : Probes which provide vast amount of info on your body. E.g 2 year warning of a heart attack, instant detection and treatment of cancer, radiation therapy (which is why Eli needs it), etc.
v2 : Red blood replacement which provides far better oxygenation ability and CO2 absorption so you can hold your breath for nearly an hour or can exercise much better. Also has amazing toxin scrubbing and processing.
v3 : The immune system is augmented. Nanobots acting as advanced White cells, T cells etc. Allows for not just a super immune system, but wirelessly transferred anti-viruses. Very quick detection of new infections, diseases and viruses and the ability to transmit an internal scan to be remotely processed and and anti-virus quickly developed and downloaded. Some neural implants can do a basic AV processing but it takes longer and takes up almost all of the processing power. Note: The nanobots are created in nanobot factories, usually embedded in the bone marrow and a couple of other points around the body, nanites (self-replicating nanobots) are NOT used due to their possible ability to go rouge.
V4 : There’s usually more than 30% of the blood as nanoblood. It also has DNA scanning facilities which, with the help of a machine that’s a bit like an MRI, allows all the cells in your body to have their DNA read.

* neuroMod – The nerual implant. See also the 10 levels of neuroMod integration. Most of the apps are aimed at level 6 or 7 levels of integration.

* bodyApps (as opposed to bodyMods) are those which use the neuroMod to alter how you move, think or interact with implants and body/bioMods the vast majority need a neuroMod and need special permissions. Kinda like when you authorise an app in Google Play, but with a lot more info about what it will actually do and not do. Especially when an app is initiating activities for the first time, like LieToMe changing the way your eyes scan someones face or the complex, stocastic movements of SpeakEazy changing how your facial muscles work to make it harder for people to detect your lies. bodyApps are more about movement and control. The Posture Pedic bodyApp is installed by default (for those with the Trev special set of mods), and goes well with the spinalTap bioMod. It makes you sit up straight, keep your head back (not in the forward head posture) etc.. Works to reduce most muscle fatigue and skeletal tension. There’s versions for running better, meditation, various martial arts. This is different to the “I know Kung-fu” part of The Matrix. It’s a neuroMod running to alter your motion control. It changes how you’d try to move.

* visionApps – As with other apps, this uses the neuroMod. You can get Public, Private and shared (group) augmented overlays. But advanced vision mods can tap into how your visual processing neurons work. You’ve got to be careful when you start changing how you perceive straight lines or other core things. It’s very easy to go into dangerous trips with reality distortions few drugs can even get close to. Whilst these days they are automatically detected and the changes reset, similar to going into a ‘preview’ mode of new monitor settings and it auto-reverting, it can be possible to get yourself stuck in a mode where you simply can’t navigate to cancel such a mode and can cause long term pshycosis. Note: You can get eyeMods which are specific eye replacements. This is what the Death Squad have. Their eyes glow red because it’s creating infrared light to help them see at night.

Some Cool Apps

NeuroTelepathy or usually just known as Telepathy is the app which lets you talk with other neruoModded humans but also with AI and modded Animals.

There’s various levels of communication. From the equivalent to IRC chat or normal speech. Very carefully controlled to pre-recorded and edited thoughts with concepts, visuals, feelings and the like to a full stream of consciousness. As you feel it or think it with only basic filtering, e.g remove most background body sensory info, anything sexual or socially inappropriate.

You can also send concepthesis concepts and more.
The Mind Meld app is the NeuroTelepathy app with a 2 (or sometimes more) way merge and no filters. Obviously named after a Vulcan mind meld.

Penfield – Emotional control
Psychology and personal mental control beyond any advanced meditator. Like the Penfield Organ in Do Androids Dream of Electric Sheep.
Gives you the ability to control your emotional state and even your thoughts with great precision. If you want.
Most people who use Dataism as their new religion give this control over to AI algorithms which can optimise their life for being the most rewarding and fullfilling, with lots of time in the state of flow. Thus providing great satisfaction beyond just being stuck in the ‘Happiness’ setting.

BabelFish – Universal translator
A core app.
Can work on both audio and visual inputs, acting like a super better version of Google Translate’s camera mode, or the conversation equivalent. Often the main way you know it’s even working is because the audio/video signal has a ‘translated’ tag added and you can toggle the translation layer and see the underlying signal before translation.

It also lets you speak or write nearly any language, if your vocal cords or hand is capable of the output. Often computer to computer digital speech and text isn’t really replicable by humans without speakerbox bodyMods (speakers instead of voice boxes) or text printing capabilities… Or usually just a digital display like eInk skin.

Lie To Me – Lie Detector
This works through all the usual input systems with some degree of accuracy against those not actively blocking it.
It hooks into the brains existing system 1 Fast triggers, but is also able to analyse with a lot more skill and accuracy than most people’s innate lie detection. It usually help focus the fovea on the person of interests likely telltale signs, mostly on their face, from their eye movements to minor muscle twitches. Because it’ll urge the user to look at certain points (they’ll want to do so), it can often be detected by others, so more stealthy modes are available, or usually people use an alternative visual input, like a nearby drone.
It uses the latest in nerualnet and other AI analysis to make you instantly a better lie detector than any unmodded person on the planet.

SpeakEasy – Lie to people / Pokerface
Usually the SpeakEazy counter app is also installed by people. It started by detecting the other persons obvious signs of using the LieToMe mod.
It then tried to intercept your tells and stop those micro-expressions. This dampening of your tells works well against muggles (unmodded people, but the weird times of dead movements then became a tell themselves.

Because such dampening is detectable the normal mode when talking to someone else that is also modded / with a Wizard Hat does the equivalent of creating white noise but with your face muscles, eyes and nearly everything else. You’ll have a sea of random micro-twitches, erratic heart rate and seem jittery to the system in a way that masks your emotions and signals by looking like you are changing emotions and sending lies and truth signals in such rapid succession it becomes meaningless.
Think of the big face of The Matrix core that Neo talks to at the end of the 3rd Matrix film, or other such particle, water or electrical based characters. There’s general form there but it’s annoying to view for a long time. Negotiators refuse to talk to people with this mode on, but those doing the last remaining bits of capitalist politics (in the non-RBE cities that aren’t New Eden) have this mode on by default.

People using the noise version often refer to it as PolkaFace, a play on poker face.

Look to the Stars – Where Am I ( Night Sky )
Look into the night sky and based on the stars know your position in the solar system (not just lat/long on Earth or Mars but anywhere out to the Oort cloud and within a +-300yr time range). The full version attempts to work out your position in the Galaxy over a +-5,000yr range and within the Galaxy. Although the full version needs a larger download and is much faster processing if you have a smart watch or even a space ship to help with the processing.

See that Key
Look at a key or any object and once you’ve seen it from enough sides you can have a 3D model generated and be able to 3D print it.
Works great when you’ve got a large drone nearby that can 3D print it for you, or are simply near a city which has fast 3D printers.

Morse code reader
Once installed this runs in the background looking for morse code signals, especially audio beeps and light flashes and will detect and convert. This allows for conversion of QS and other morse codes into their general understanding.
It will also detect binary, and other basic protocols by default.
This is often hooked up to people’s flash light fingers and it can be fun to see two kids running around talking via their fingers flashing, although there’s more advanced transmission protocols used for tapping and vibrations, so someone just tapping on a table or tapping on their friends hand whilst holding hands, or leg when cuddling can be a form of talking behind other people’s backs.
As with many of these things, there’s an ever evolving war of encryption and decryption, kids using a greater variety of ciphers, although the generic decryption tools and AI decryptions make basic changes to the ciphers easier and easier to break. Although the normal child / teacher empowerment dynamic is a lot more like that of a respected mentor so is rather different to normal school setups.

The voiceBox bioMod, allowing for a built in speaker, whilst not nearly as popular as light fingers is often used to transmit on high frequencies most humans can’t speak at and usually can’t hear without hearing bioMods but is often used by cohesive groups that are modded. But nearly all people with such bioMods also have the neuralMod implant so just communicate via standard encrypted wireless telepathy.

The development of the morse code app is often used as a standard example of app development of its kind. The first developer thought it would be cool to read morse code like they do in all the movies, without actually having to learn morse code so worked out how to hook into the various pattern recognition systems of the brain and neuroMod. It’s usually a combination of system 0 (basic visual neural detection of lines, shapes and time repeating patterns) plus the first order, system 1 (fast) processes which usually detects the flashes. By buffering a few seconds of input on a background thread that’s analysing the signals it can try and find anything that looks like morse code. Once detected it can analyse the input with greater focus (be it visual, audio, kinesetic or even smell) do more specific optimisations around the particular signals ‘hand’ (timing patterns, be they miliseconds in duration or minutes) and do more noise reduction, context analysis (does QSR mean a general morse code short hand, or just the 3 letters) etc…
Special patterns, like SOS are extra highlighted.

It started as more of a gimmic than anything, but others took the code and made it easily extensible so new filters could be added, new algorithms detected. More hooks into different areas, like concepthesis messaging (concepts and learnings) available as sound or touch for those who don’t have normal wireless internet to the brain enabled, usually enabled at museums or art galleries as a form of disabled / fallback support.
It can even detect messaging shown via agumented reality or as simulated external sensory input (sound, touch, vibration, smell, test) during immersive VR sessions, etc..

Then there’s a whole host of different detection algorithms, hooks into apps like the more powerful generic decryption systems. These can offload processing amongst a neruoMesh (group of other people with neruoMods), allow for joint detection of inputs (e.g thousands of people worldwide getting small pieces of the puzzle) and of course the ability to scan over very long time periods, like years not seconds, but most of those are all rarely used addons mostly done for fun, although some cool detection of earthquakes by peope who were meditating was possible on a neuroMesh. On the direction vibration sense some people will augment their heart beats to act as a morse code, detectable by their partner just by holding their hands when fairly quiet. Although heart rate detection of others is fairly easy at a greater distance with some of the infrared eye bioMods or tuned electricial EM field detectors, be they internal to the person (thus requiring a lot of work to cancel out the detection of their own nervous system’s EM activity) or external, like built into the walls and ceilings, usually trying to focus on bioelectric signals not normal wireless transmissions.

4 Versions of the Olympics
#1 – The existing normal Olympics. Healthy and unaltered. No drugs, nothing. With #1D being the disabled Olympics, although that’s a LOT rarer given most people opt for replacement grown limbs or even end up in the #3 tech enhanced version with bionic limbs that were better than what they used to have. Actually it was because of dealing with disabilities we developed such good bionics.
#2 – Drug enhanced – Humans using drugs and other general enhancements, but nothing we’d consider active or passive technology.
#3 – Tech enhanced / Cybernetic – People with nanoblood, implants, bionic limbs and many of those beyond the Kubler cascade
#4 – An android only version – Mostly for robots only. Although there’s occasionally matches between humans and Androids, e.g the Robocup Challenge style Soccer against humans and Robots. Although usually only full cybernetic enhanced people have any hope against even reasonably well optimised robots. Often the Androids will zoom on their equivalent of roller skates instead of pumping their legs, or will have very different forms of locomotion and there’s some surprising ways that the genetic algorithms for say long distance Javelin or high jump can create amazing robots. Few people even consider it the same sport, but then, the androids didn’t really see much to sport. The more interesting robots are the ones that attempt to be able to beat the best enhanced human in each area but without replacing parts. So being better than the best human at not just throwing and jumping but shooting and running and playing sports and swimming. To be good at all means some very interesting trade-offs and engineering feats.

Kubler Cascade

The Kubler Cascade is when there is some sort of driving force in a persons life which makes them want to be more augmented, e.g wanting to be the best athlete or fastest tech head they can be and in doing so quickly jump from the generally acceptable 40% range to the 80-90% level of augmentation. (I’m assuming that) It’s harder to go beyond that level due to technological difficulties and it’s usually easier to jump from that level to being digitised and having your consciousness running completely inside a computer. However, usually it’s a different force or pressure to make someone want to be digitised.

The other point to highlight is the fact that most people are fine with the up to 40% augmentation. This usually involves neural implants, basic genetic corrections, enough nanobots to ensure they are healthy and have an abundance of sensors to know if there’s anything wrong with them.
But when you apply a competitive force and you start replacing your legs with fully technical ones which allow you to run over 100km/h to do that with anything more than a quick burst you need to feed them with enough energy, so you need to up your nanoblood levels, replace your heart, become a chest breather and replace your stomach.

Obviously people augmented people in the tech enhanced version of the Olympics are the most susceptible to this.

Often the Kubler cascade is also defined as the point before passing you are still considered a normal homo-sapiens, but after which you are now classified as something else. Some people use the popularised term homo-deus or God like human. Others use De novo sapiens or just novo-sapiens meaning anew humans. Or the more pretentious French version Nouveau-Sapiens.

10 Levels of neuroMod Integration
There’s actually 7 main levels of neuroMod integration.
Once you go past around level 8.4 of integration there’s issues with being able to reverse the process and remove the neuroMod. Too much of your brain has been replaced. Hence most people are between level 7 and level 8.2 integration.

Level 0 – Nothing. You are an unmodded. A muggle.
Level 1 – The seed has been injected and there’s a nanofactory running but it hasn’t connected with anything.
Symptoms: Some light tingling at the site of the injection and maybe a tiny bit of discomfort where the neuroMod factory implant has an injection site poking through the blood/brain barrier.

Level 2 – The neuroMod is just beginning to integrate with some nearby neurons. It’s ensuring compatibility, ensuring the body doesn’t reject it.
Symptoms: During this time there might be small, barely noticeable glitches. Things like unexpected memories. But it can sometimes also trigger an out of body experience.

Level 3 – The neuroMod now has communication with the Internet via the wireless chips in your collarbone. It’s also creating main pathways to the important parts of your brain, the motor cortex, amygdala, frontal cortex all along between your optical nerves and visual cortex and it is generating more of the main highway infrastructure. The main tree branches.
Symptoms: There’s the motor cortex and sensory cortex. You could have issues where the sensation is there but you can’t move or vice versa. This would cause some weird lack of Proprioception, the feedback loop of going to move something and then feeling the texture, weight and other senses that let you know about that object.
The weirdest is when it mutes the channel about reliability / probability of sensory input, everything feels uncertain. You don’t know the certainty of what you are perceiving, it can get very weird.

Level 4 – Initial visual integration. You’ve got visual overlays as the neuroMod is now interrupting the optical nerve.
Symptoms: At first you get vision that seems empty (no signal) and the brain fills it up with imagined creatures or shapes ( like the visual aura I get before having a migraine or Oliver Sack’s talk about Charles Bonnet syndrome ), or distorted and weird. Although by the end you have Augmented Reality and basic close your eyes Virtual Reality. You interact with the overlay using basic eye tracking gestures (seeing as that’s controlled by the lower level lizard brain not your motor cortex… which is cool).
Also, the midbrain controls your voluntary eye movements. What happens when it feels like the neuromod forces your eyes to view something? I’m guessing there has to be a thought like request or recommendation before taking on an eye movement without it feels weird, but it’d be simple and subtle. Could be a weird sensation to make use of esp during AR gaming.

Level 5 – Full Motor cortex, and emotional integration. It can now change the way you move and how you feel.
Initial sensory based Life Stream recording happens here as all the main sensory input is being recorded and the global workspace (conscious thought) has been decoded to enough of an extent.
Symptoms: Sometimes you’ll have weird twitches as it triggers some muscle groups and sudden emotional outbursts or numbness. You’ll also likely get some brief out of body experiences.

Level 6 – Thought manipulation. No longer just reading your thoughts, able to change and manipulate them. Able to start mapping your memories.
Symptoms: <Insert>

Level 7 – FULL Integration. You can have Full Virtual Reality (FVR), Matrix style. An immersion so strong that without certain restraints and your memories you might not know you are in a simulation.
By now you can start to think about 3x faster than normal and sleep only 2.5hrs a night. Your a Wizard Harry!
Symptoms: <Insert>

Advanced neuroMod Levels
Level 8 – Technomancy. This is for the speed freaks and involves using the neuroMod along with a whole bunch of enhancements to increase the speed of your thought by another 5x, so 15x faster than normal (or 23ms instead of 350ms response times). With certain very special capabilities, like bullet time, being done as ASIC like dedicated hardware in order to get something decent out of 350ms of bullet flight time you’d want 5ms time slices for 70 frames of action and reaction, so it’d feel like about 3s.
This is the level that the technotopians are required to be at, but also the negotiators as they have the Bullet Time mod.
Symptoms: Your able to process things so fast that conversations with people of a lower speed level aren’t possible in realtime thought. You have to buffer the input and output. You also find that Internet latencies become noticeable so huddling near data centers or physically being near the groups of people you are telepathically talking with becomes important. Also, the amount you sleep will have been drastically reduced to about 30mins a day due to advanced neurochemical cleansing.

Level 9 – ? Are you human, cyborg or what?

Level 10 – You’ve not longer got a biological brain, it’s completely replaced. This hasn’t been achieved… Yet. But people are working on level 9 and 10.

Level 10+ would be a fully digitised consciousness.

Initially complied by Michael Kubler on the 19th of October 2018 from a variety of his ideas for the Book of New Eden novel. Inspired to post this by Geoff Kwitko’s Live Stream talking about his interest in SciFi.