This is my personal notebook where I dump thoughts and notes for my future self.
**I make no guarantees about the accuracy of anything in here. Read at your own
risk!**
[TOC]
# August 2023
## 2023-08-21
First day of orientation as a PhD student. Here we go again, back to school one
final time.
Figured out the campus wifi despite the Linux jankery. Had to:
1. Register as a special device, the laptop registration link redirected to
nothing useful.
2. Use `nm-connection-editor` from `gnome-network-manager` to edit the
connection and manually set up the WPA2+PEAP+user/pass for the connection.
Finally fixed the dulwich errors from hg-git. Since I'm using Debian's hg now
instead of building from source, I needed to install dulwich somewhere that the
system python can find it. I almost installed `python3-pip` to do that, but
then realized I could just install `python3-dulwich` and be done. Cool.
## 2023-08-22
Day 2. Lots of getting talked at, meeting with advisors, etc. Still trying to
get all the moving parts settled down — hopefully it'll be a lot clearer once
classes and rotations start (though I'm sure it'll be busy in a different way).
First time trying the Ann Arbor bus system, the bus was 20 minutes late. This
bodes well.
## 2023-08-23
Got a presentation about things to do/see/eat in Ann Arbor. Lots of things to
try.
Advice from existing student panel:
* Keep in touch and make connections with folks in your cohort/community who are
going through the same stuff as you.
* Rotate with anyone your want, don't let anyone tell you not to.
* Don't forget to have a life. Don't spend every weekend in the lab.
* Don't compare yourself to each other. Lots of variety in the incoming people.
* Ask for help.
What to ask/think about when looking for a lab:
* Do you have funding to support me?
* Do you have plans to stay at the university?
* Talk to current and previous students (and where those ended up).
* Talk to multiple students (people have different experiences).
* Work ethic.
* Mentoring style.
* What are your expectations of me?
* Priorities toward students (high/low?).
* How did they handle difficult situations in the lab?
* Find a good PI, because they're the one that will be there the entire time
(students/etc can and will leave).
* Tell them what you want to do (e.g. go to conferences), and based on their
responses you'll know whether they're a good fit.
* Switching labs is possible (not ideal though).
* Listen when people tell you a place sucks.
* Alumni are great because they don't have the power dynamic to worry about and
can actually be honest.
* You can change rotations if you really need to, even if it's the second week
in.
* Ask how many slots there are vs how many rotations they're taking to judge how
competitive joining the lab will be.
## 2023-08-24
Did the Python pre-test so I don't have to take the introductory programming
class.
Did more online training.
More talks. Lots of meeting people after.
## 2023-08-25
Lost power last night because of the storms and DTE says I won't get it back til
tomorrow. This is… not great.
Doing more trainings and paperwork at a school building so I can plug in my
laptop.
## 2023-08-28
Turns out I didn't get power until *last night*, i.e. three of God's own full
days after it went out. Ann Arbor is… not feeling great right now.
Also, yesterday around 14:00 the university's network went down, entirely.
Unable to get online from the wifi, and any services hosted on the university
network are down, which is… not great. Of course, many of the services used
(e.g. Canvas) are third-party services not hosted on the university network.
Except they all use the university SSO (because Security™), which *is* hosted on
the network. So you can only use them if you're already logged in. And of
course sessions expire super quickly (because Security™) so, in summary: lol.
Starting to use my HP-15C clone(s) from Swissmicros again. Was getting
misleading results when trying to do some stuff with logarithms: it seemed like
it only had two digits of precision. But after some digging around I realized
I must have configured the *display* precision to 2 SD's at some point in the
past and forgotten about it. The full precision is there, it was only rounding
to display. I set it to 6 to avoid this, but also the `PREFIX` key will show
the full precision temporarily.
Figured out (again) how to program the calculator to do logs of arbitrary bases.
Essentially:
logₓ(y) = ln(y)/ln(x)
stack
LBL C x y
LN x ln(y)
x<->y ln(y) x
LN ln(y) ln(x)
÷ ln(y)/ln(x) = logₓ(y)
RTN
Trying out syncthing as a Dropbox replacement. Installing:
sudo apt install syncthing
systemctl --user enable syncthing.service
systemctl --user start syncthing.service
Then <https://localhost:8384> to access the admin.
Got my budget scripts working and synced via syncthing (also shaved a couple of
yaks by making scripts to archive/create new hosts while I was at it). Seems to
work okay at the moment. Will gradually transition other stuff over time.
Going to spend some time learning about Nextflow while I wait to hear from
rotation folks. Nextflow is basically a DAG, where:
* Edges are FIFO queues (Nextflow calls them "channels")
* Vertices are things that consume input from their channels and produce output (Nextflow calls them "processes").
There are two types of channels. First: queue channels: asynchronous FIFO queues. Examples:
# emits sequence of given values
ch = Channel.of(1, 1, 2, 3, 5, 8)
# emits a single file path (queue of size 1)
ch = Channel.fromPath('data/one-single-file.txt')
# emits multiple file paths
ch = Channel.fromPath('data/*.txt')
Value channels: like queue channels, but just emit the same value over and over.
Basically `(constantly val)`.
Processes: basically stages of a pipeline. Take input and output definitions,
plus something to run (e.g. `shell`). Example after a bit of poking around:
// allows you to define processes to be used for modular libraries
nextflow.enable.dsl = 2
workflow {
ids = Channel.fromPath('data/ids.txt') // single-item channel
chunksize = Channel.value(1000) // (constantly 1000) but will only ever be used once here
// The process below produces a list of outputs. It will only ever be
// run once, but nextflow doesn't know that -- you could potentially
// have a process run multiple times that each produces a list. So by
// default it groups all the outputs into a single emitted value. But
// here we want to flatted [[aa ab ac ad ae]] into [aa ab ac ad ae].
batched_ids = split_ids(ids, chunksize) | flatten
batched_ids.view() // .view() doesn't consume, good for debugging
result = reverse(batched_ids)
result.view()
}
process split_ids {
input:
path(ids)
val(chunksize)
output:
file('batch-*')
shell:
"""
split -l !{chunksize} !{ids} batch-
"""
}
process reverse {
input:
path(batch_file)
output:
file('result')
shell:
"""
tac !{batch_file} > result
"""
}
Nextflow seems to have the concept of a "run name", i.e. an identifier for
a particular run. It creates a `work/` directory with the output files, but
*also* seems to splat out a bunch of hidden `.nextflow/` and `.nextflow.log.*`
files in the current directory. `nextflow clean` removes `work` but not the
hidden files.
Can run with some basic reporting with some extra flags:
nextflow run example.nf -with-report report.html -with-dag graph.svg
Of course there appears to be some backwards-incompatible jank in the language
already. Reading through <https://www.nextflow.io/blog/2020/dsl2-is-here.html>
shows minor syntax changes I guess I'll need to be aware of when looking at
things a few years old.
Putting some things on the TODO list for learning mroe about nextflow:
* <https://github.com/seqeralabs/nextflow-tutorial> (uses old DSL)
* <https://carpentries-incubator.github.io/workflows-nextflow/index.html>
I also need to get some basic scratch VM infrastructure set up with qemu/vagrant
and Ansible so I can test out Nextflow/pipeline stuff without polluting my own
machines and/or any servers I eventually get access to with random testing
garbage. Maybe I'll put that on tomorrow's list too.
## 2023-08-29
First BIOSTAT-521 class. This one seems like it's going to be quite easy, but
given I have a crazily-hard class and lab rotation work to do, I think that's
fine with me. It *does* use R, which will be a good excuse to poke around at
that since it's used so heavily in the industry.
First lab meeting as well. It was good to meet everyone in person. Talked
about scheduling stuff and overall gist of my project, though we can't move
forward directly right now while the university network is still hosed. But
I did get some more information about things I'll want to look into:
* Snakemake (not Nextflow, oh well).
* Singularity containers (should probably read the UNIX book section on
containerization first).
* I still want to get Vagrant/libvirt/qemu/Ansible working to make scratch envs.
Going to meet again on Thursday to see if the network is available. If so I'll
dive in more then — until then I'll just poke around learning that stuff because
I know I'll need it eventually.
Don't have a lot of time before next class, but I went ahead and installed
Snakemake highlighting/etc for Vim.
## 2023-08-30
HG545 this morning. Class was mostly about how to succeed in the class.
Still trying to decide on how I want to take notes while I'm here. I read the
Zettelkasten book and was considering trying that. But after poking around at
it I'm not sure I like the overhead of having to link everything up all the
time. I tried creating some notes while studying and it was a pain to have to
try to link everything to something else. Sometimes I want to just jot down
something without worrying about its place in the graph. I think I might end up
going with this current format (stream-of-consciousness .plan file style notes)
for everything, but taking a few things from Zettelkasten than might help:
* Take fleeting notes as I read.
* Turn fleeting notes into permanent ones, but as text in my .plan instead of
linked entries in some other system.
That seems like something I might actually *do*, and hopefully with grep it'll
be good enough for what I need.
Installing JabRef to try as a reference manager. Zotero looks nicer (no stupid
flat UI) but syncing the DB requires sending it to their web thing. JabRef
seems to use a plain text file so I can probably just sync it with syncthing and
deal with conflicts manually. Spent some time adding a couple of papers to it.
Not sure it's great (it got the info wrong for 2/3 papers) but I guess that's
just typical open source jankery.
Apparently you can just `C-v` a DOI into JabRef and it'll import it. Hard to
discover, but seems to work okay. JabRef is complaining about capital letters
in the titles but I'll figure out that jankery later. At least I've got
something for now.
Had some wonkiness with my Syncthing budget stuff, but I think I just forgot to
reeval the location on my desktop. Will poke around more if anything else seems
to break.
Watched some snakemake videos and read through their paper. This smells a lot
more academic than Nextflow did, which is a little worrying. I'm sure it'll be
fine in the end though.
Send off the rest of my VA paperwork so things can get moving on that side.
Read the ULSAH section on containers to get a high-level overview. Started
looking into Singularity and it's already looking spicy. Apparently the project
forked a couple of years ago and there are now two competing versions? Great.
Also you have to install it from source, which requires installing Golang.
I thought I was free of Rob Pike's Googly Tendrils but I guess I never will be.
Installed Go, built Singularity. At least it installs to a prefix
(`/opt/singularity`), so I can remove it easily if I want.
Poked around a little to make sure it's working, e.g.:
singularity pull docker://debian:bookwork-slim
singularity shell debian_bookworm-slim.sif
Seems to be working as far as I can tell.
Also installing snakemake. Using pip with a venv for now even though the
documentation tries to convince you not to. If anything breaks I can revisit
it, but for now it's probably fine to go through some tutorials without pulling
in some giant slab of junk.
Started going through the Snakemake tutorial.
> Since the rule has multiple input files, Snakemake will concatenate them,
> separated by a whitespace [sic]
Oh boy.
Realized I'd need to install a pile of stuff to get through the tutorial,
decided to pause and shave the qemu yak first so I can do this without dumping
a ton of stuff on my laptop. So many yaks.
Shaved the qemu yak, now I've got a reliable VM setup. Committed the
instructions and a tiny script to a `vms` repo so I don't have to relearn this
again.
With that out of the way, installed Snakemake and all the prereqs from their
tutorial on the VM with wild abandon. Now I can *actually* do the tutorial.
The simple tutorial was straightforward for the most part, but for this:
rule bcftools_call:
input:
fa="data/genome.fa",
bam=expand("sorted_reads/{sample}.bam", sample=SAMPLES),
bai=expand("sorted_reads/{sample}.bam.bai", sample=SAMPLES)
output:
"calls/all.vcf"
shell:
"bcftools mpileup -f {input.fa} {input.bam}"
" | bcftools call -mv - > {output}"
It's not clear how the expanded input lists are ordered. Are they guaranteed to
always produce the same order given the same input list?
Finally realized I could put my pubkeys online so I can just `curl
https://stevelosh.com/pubkeys >> ~/.ssh/authorized_keys`. No idea why I never
though of this before now. Also updated my site now that I've moved. Of course
everything just worked even though I haven't touched the site in months, because
it's written in Common Lisp which never changes. It's so nice to not have to
work with constantly-breaking shit.
Installed R and RStudio for tomorrow's class. Base R is `r-base` in Debian.
Unfortunately Debian 12 has only been out for a few months and the RStudio
Debian package only supports 11, but apparently the `deb` one for Ubuntu 22
works so I guess I'll just yolo it with that one for now.
Reading for tomorrow's BS521 class: chapter 1 of OpenIntro Statistics (probably
also want to real Holm's stats book again as a refresher). All pretty basic so
far.
## 2023-08-31
BS521 this morning. Still pretty straightforward.
Going to create a separate note repo for lab notes, so I can take those
privately. I'll still put public notes here, but for certain things (e.g. DB
table names, etc) I don't want to have to worry about which are okay to make
public.
Going to port the abortive attempt at a Zettelkasten into here in no particular
order (thank god for `grep`). Should have done this earlier but it's been
a hectic couple of days. Should I use heading more in this .plan for folding
purposes? Maybe. I still need to finish going through my fleeting notes that
I've got so far and getting them in here. I also should have already done this,
but I'll do that this weekend when I have some free time and try not to let it
pile up so much in the future.
### Short-Read Sequencing
Short-read sequencing aims to sequence long stretches of DNA (or RNA) by first
fragmenting them into small pieces, sequencing those pieces separately (usually
in parallel, for speed), then reassembling the fragments into contigs
computationally.
There are several forms of short-read sequencing, but the most popular by far
today is Illumina's sequencing by synthesis approach.
Short-read sequencing is often used for whole-genome sequencing.
The cost of sequencing, especially short-read sequencing, has fallen
*dramatically* in the past 20 years, which has started to enable its use as
a clinical tool. It was impractical to sequence someone's genome to learn about
their health when it cost $1 billion, but it is much more reasonable to do it
when it costs $1,000.
### Whole-Genome Sequencing
Whole-genome sequencing refers to sequencing an organism's entire genome,
instead of just a limited part of it.
This has the benefit of letting you get data on *all* of the genome at once,
without having to know what you're looking for in advance. But if you *do* know
what you're interested in, it can be much faster/cheaper to sequence just that
part (or can give you deeper coverage for the same cost). Like many things in
science and engineering, it's a tradeoff.
### Fragmentation
When sequencing DNA or RNA with short-read sequencing like Illumina, the strands
of nucleic acid need to be fragmented into shorter pieces. This usually happens
in one of a few ways:
* Mechanical fragmentation by forcing it through a small passage
* Enzymatic fragmentation
* Sonication
### Sequencing Alternatives
Although sequencing costs have fallen dramatically in the past 20 years, they
are not completely free. There are several alternatives to sequencing that are
still used for a number of reasons:
* They can provide the information more cheaply than sequencing.
* They can provide much deeper coverage than would be practical to get by sequencing.
* They can be done much faster than sequencing, allowing for rapid diagnostic use.
Some examples of alternatives are:
* Microarrays
* Nanostring
* qPCR
* Optical mapping
### Post-Transcriptional Modification
After RNA is transcribed from DNA, but before it goes on to do its actual job
(being translated into protein, or doing something on its own), it is sometimes
modified. There are a number of modifications that happen, some of the most
common are:
* Splicing out introns.
* Polyadenalation of the 3' tail.
* Adding a 5' cap.
### Post-Translational Modification
After RNA is translated into protein, but before the protein begins to do its
intended work, it is sometimes modified. Modifications come in many forms and
can affect the terminal ends or one of the amino acid side chains. One common
modification is phosphorylation.
### Chromatin
Chromatin structure varies in the genome. Chromatin that is uncoiled and
available for work is euchromatin (eu = useful). Chromatin that is tightly
coiled and not accessible for transcription is called heterochromatin.
Heterochromatin has two sub-types of its own:
* Facultative heterochromatin can temporarily become euchromatic.
* Constitutive heterochromatin is always condensed, and is usually found in
repetitive sections of the genome.
[ATAC-Seq](https://en.wikipedia.org/wiki/ATAC-seq): "Assay for
Transposase-Accessible Chromatin using sequencing" is used to determine which
parts of the genome's chromatin are accessible (and, implicitly, which are not).
It uses a mutant, hyperactive transposase to insert sequencing adapters
*everywhere* it can inside the genome. Those tagged fragments are then
extracted and sequenced.
[ChIP-seq](https://en.wikipedia.org/wiki/ChIP_sequencing):
"Chromatin-immunoprecipitation sequencing" is used to determine which parts of
genome particular proteins are interacting with, e.g. where a particular
transcription factor is binding to in the genome. The steps are roughly:
1. Crosslink the DNA and proteins to lock them together, so the protein can't
unbind from the DNA (usually using formaldehyde).
2. [Fragment](n/fragmentation.md) the DNA into small (~500bp) fragments.
3. Use immunoprecipitation to extract only the proteins you are interested in
(antibodies that bind to your protein of interest).
4. Unlink the DNA and proteins, extract the DNA, and sequence it (e.g. with
Illumina).
The resulting reads will be regions where the protein of interest was originally
bound to the DNA, allowing you to see (roughly) where a particular protein
interacts with the genome.
### Homopolymer
"Homopolymer" when used in the context of sequencing means a sequence of the
same base repeated, e.g. `AAAAA` or `CCC`.
Many sequencing technologies have trouble with these kinds of sequences. For
example, Oxford Nanopore sequencing uses the properties of the bases currently
inside the pore to determine the bases that enter/leave. But for homopolymers
it's hard to tell when one enters and another leaves if they're completely
uniform along the entire pore.
### Long-read sequencing
Long-read sequencing is a relatively new branch of technology that allows for
sequencing much longer reads than e.g. Illumina. This is useful because many
parts of many genomes have regions that are difficult for short-read sequencing
to deal with, e.g. repetitive elements and high-level structural variation.
Long-read sequencing can help assemble those sequences which would be difficult
to piece together from short reads.
### Oxford Nanopore Technologies
A long-read sequencing technology that can be used to directly read a strand of
DNA using a molecular pore that produces current as DNA passes through it. The
variations in current can be recorded and processed to determine the DNA
sequence being passed through.
---,
\
\
=============== DNA helix
/
/ ssDNA (unwound)
/
( | )
( | ) motor protein
( | )
[ | ]
[ | ] pore for sensing DNA
[ | ]
[ | ]
[ | ]
______[ | ]__________________________
|
|
vvvvv DNA is fed through the pore by the motor protein and current (picoamps) measured
ONT units are available as small devices that connect to a USB port on a laptop,
which makes them much more portable than other sequencers (though the library
prep equipment is not as portable, yet).
### 10X Genomics Linked Reads
A form of synthetic long-read sequencing from 10X. Long DNA fragments were
loaded into gel beads with barcodes. They were then fragmented and barcoded,
then sequenced on a normal sequencer. The barcodes allowed you to know which
short reads came from the same longer fragment (i.e. were "linked"), which helps
reconstruct the longer fragments from the short ones.
We discontinued this product while I was working at 10X. RIP.
### Non-uniform genotypes
An organism is a **mosaic** if it has cells with different genomes that stem from
a somatic mutation in one of the ancestral cells, which then divided and
resulted in a chunk of the organism having the mutation while the rest does not.
An organism is a **chimera** when it has cells with different genomes that are
caused by multiple zygotes combining early in development. The resulting
organism will have cells with completely different genomes depending on which
zygote they trace their ancestry back to.
### Kaelin 2017
Read this paper from intro materials. Main thrust was about broad vs narrow
claims in papers.
Scientific papers make and justify claims using evidence. Papers vary on how
many claims they make, and how well-supported each of those claims is with
evidence in the paper.
Because scientists have limited time and papers must be a finite size, there is
almost always a tradeoff between making a broad set or a narrow set of claims:
more claims will generally mean less evidence to support each claim.
The author claims that in biomedical research there is a growing trend toward
requiring papers with broader and more sweeping claims if you want to get
published. Reviewers want more and more claims, more "translatability" to the
real world. It is no longer enough to notice and document something is
happening, now you must also propose *why* it happens and experiment to
demonstrate this mechanism.
Requiring authors of scientific papers to make broader and broader claims in
order to get published has several (likely unintended) effects.
First, by adding more and more claims to a paper, each claim will likely have
less individual support behind it. Instead of a few claims supported by many
pieces of evidence:
------- -----
||||| |||
papers will end up with a collection of claims, each balancing precariously on
narrow support:
--------- ----- --- --- ----- ----
| | | | | | ||
This harms reproducibility because claims that are not supported by a robust set
of evidence are less likely to reproduce successfully.
Broad claims are also often harder for peer reviewers to deal will -- they often
need to be an expert in multiple fields just to be able to evaluate the paper.
### Recombination
Recombination is a step that happens during meiosis where homologous chromosomes
cross over pieces of themselves, effectively swapping random chunks of their
sequences. This provides much more genetic variation than random segregation of
chromosomes alone, allowing sexual reproduction to explore more of the genetic
space more quickly.
The number of recombination events (crossovers) per chromosome is random, but is
usually relatively low (3-5 per chromosome).
# September 2023
## 2023-09-01
HG545. Looked over the slides last night and was a little worried, but felt
okay after the lecture for the most part. Still a few things I need to look up
and I do still need to get my fleeting notes into this, but I feel okay.
Continuing the Snakemake tutorial.
Threads can be specified for a given job with `threads: 8`, and you need to
propagate that to the command yourself with `{threads}`. Will be scaled down if
run with fewer cores than threads, otherwise will wait until that many are
available.
Snakemake has some support for noticing log files, but it seems like you have to
manually create them yourself? This seems… tedious?
rule bwa_map:
input:
"data/genome.fa",
lambda wc: SAMPLES[wc.sample]
output:
"mapped_reads/{sample}.bam"
threads: 8
params:
rg=r"@RG\tID:{sample}\tSM:{sample}"
log: "logs/bwa_map/{sample}.log"
shell:
"("
"bwa mem -R '{params.rg}' -t {threads} {input}"
" | samtools view -Sb - > {output}"
") >{log} 2>&1"
Do I really have to wrap everything in `(…) >{log} 2>&1` by hand myself?
You can get a summary of file provenance with `snakemake --summary`. The output
is a TSV, so I went down a rathole of pretty-printing TSVs and eventually found
that `| column -s $'\t' -t` works (mnemonic: `s$tt`). I love how every UNIX
program gets to invent its own bespoke command line interface for specifying
special characters. Really great.
Can mark outputs as `temp()` and `protected()`, which is nice.
Need to install singularity *inside* my VM:
# Ensure repositories are up-to-date
sudo apt-get update
# Install debian packages for dependencies
sudo apt-get install -y \
wget \
build-essential \
libseccomp-dev \
libglib2.0-dev \
pkg-config \
squashfs-tools \
cryptsetup \
runc
# Install Golang
export VERSION=1.21.0 OS=linux ARCH=amd64 && \
wget https://dl.google.com/go/go$VERSION.$OS-$ARCH.tar.gz && \
sudo tar -C /usr/local -xzvf go$VERSION.$OS-$ARCH.tar.gz && \
rm go$VERSION.$OS-$ARCH.tar.gz
echo 'export PATH=/usr/local/go/bin:$PATH' >> ~/.bashrc && \
source ~/.bashrc
# Install Singularity
export VERSION=3.11.4 && \
wget https://github.com/sylabs/singularity/releases/download/v${VERSION}/singularity-ce-${VERSION}.tar.gz && \
tar -xzf singularity-ce-${VERSION}.tar.gz && \
cd singularity-ce-${VERSION}
./mconfig && \
make -C builddir && \
sudo make -C builddir install
## 2023-09-02
It is time to shave the LaTeX yak again. Installed it with `texlive-latex-base`
to start, we'll see if I need to add some more crud in later. Going to go
through some guides for now.
Going to note some things to remember. Skeleton of document:
\documentclass{article}
\begin{document}
Basic text.
\end{document}
Math:
Inline math $y = 3 \sin x$ example.
Block equation:
\[
y = 3 \sin x
\]
With reference:
\begin{equation}\label{equa}
y' = 3 \cos x
\end{equation}
refer to it by label, e.g. equation (\ref{equa}).
More complicated: $x^2$ and $x^{2+\alpha}$ and $y_{n+1}$.
Verbatim:
Verbatim text: \verb"$x^{2+\alpha}$". Delimiter can be anything ala sed,
\verb_%%&_ or \verb+$$+.
Must escape special characters \&, \$, \%, \_, \{, \}, and \#.
\begin{verbatim}
A whole verbatim region.
(defun square (x)
(* x x))
\end{verbatim}
Comments:
Comments exist. % This is a comment.
Type styles:
Shapes:
\textup{Upright}
\textit{Italic}
\textsl{Slanted}
\textsc{Small}
Series (weight):
\textmd{Medium}
\textbf{Boldface}
Families:
\textrm{Roman}
\textsf{Sans}
\texttt{Typewriter}
Emphasis:
\emph{Never} do Foo!
"Environments" are sections that are treated differently, made with `\begin{…}`
and `\end{…}`.
Lists:
Unordered list:
\begin{itemize}
\item Foo
\item Bar
\item Baz
\end{itemize}
Ordered list:
\begin{enumerate}
\item One
\item Two
\item Three
\end{enumerate}
Customizable labels:
\begin{description}
\item[Rule 1.] Foo
\item[Rule 2.] Bar
\item[Rule 3.] Baz
\end{description}
Sizes (note the brace comes BEFORE the command!):
{\Huge Huge}
{\huge huge}
{\LARGE LARGE}
{\Large Large}
{\large large}
{\normalsize normalsize}
{\small small}
{\footnotesize footnotesize}
{\scriptsize scriptsize}
{\tiny tiny}
Centering:
\begin{center}
{\large\textbf{Assignment 1}}\\% The \\ linebreaks.
Steve Losh\\
BS521
\end{center}
Example table.
\begin{tabular}{l|rc} % lrc = cols should be left, right, centered, pipe for vertical line
Name & Mark & Grade \\
\hline\hline
Foo & 99 & A+ \\
Bar & 51 & C \\
Baz & 5 & F
\end{tabular}
Colspan with multicolumn command.
\begin{tabular}{|l||r|r|}
\hline
& \multicolumn{2}{c|}{Grades} \\
\cline{2-3}
Name & Class 1 & Class 2 \\
\hline\hline
Foo & 99 & 88 \\
Bar & 51 & 65 \\
Baz & 5 & 58 \\
\hline
\end{tabular}
Full example, with referencing and caption, e.g. `Table~\ref{tab:a} on page~\pageref{tab:a}`.
% b = try to put at Bottom. Also t top, h here, p separate page.
% Can do multiple in order of preference.
% [!t] ! = try harder
\begin{table}[b]
\begin{center}
\caption{An Example Table}
\label{tab:a}
\begin{tabular}{lr}
Name & Value \\
\hline
Foo & 1.0 \\
Bar & 15.9 \\
Baz & 6.2
\end{tabular}
% \caption{Caption at the end works too.}
\end{center}
\end{table}
Sections:
\section{Some section} % includes numbering
\subsection{Some subsection}
\section*{Some section} % no numbering
\subsection*{Some subsection}
Quotation marks (hilarious):
`Single quoted'
``Double quoted''
Change overall text size (simple):
\documentclass[11pt]{article} % only valid sizes are 10/11/12.
Palatino instead of Computer Modern:
\usepackage{mathpazo}
\linespread{1.05} % needs more leading (space between lines)
\usepackage[T1]{fontenc}
Vimtex stuff:
* Close thing with `]]` in insert mode.
Got about that far, which was enough to start my BIOSTAT-521 homework. Will dig
in again later, but it's nice to be able to use it for something real to
practice.
Puttered around a bit looking at other fonts, but didn't find anything new or
interesting.
## 2023-09-03
Spent most of today getting the reading room in my apartment ready. Went to
IKEA, looked around a lot and got some ideas, picked up a chair and bookshelf
for my still-boxed unread books. Not the most productive Sunday, but sitting in
my chair and reading at night felt good.
## 2023-09-04
Continuing the EdX genetics course over breakfast.
## 2023-09-05
BIOSTAT-521 and BIOINF-500 classes this morning.
Going to spend my non-class time looking into Unicycler today (and taking care
of paperwork if anything that needs my attention crops up) and Bandage to
visualize the results.
Grabbed the container from StaPH-B and tried to figure out where the executable
is. I think it's `/unicycler/unicycler-runner.py`. Grabbed the sample data
from the Unicycler repo and got something running:
singularity exec containers/unicycler.sif \
/unicycler/unicycler-runner.py \
--short1 sample_data/short_reads_1.fastq.gz \
--short2 sample_data/short_reads_2.fastq.gz \
--out assembly/
Pretty straightforward so far. The result directory contains a log and a bunch
of GFA files. GFA apparently stands for [Graphical Fragment
Assembly](https://gfa-spec.github.io/GFA-spec/GFA1.html), but I think it's
"graphical" in the "directed/undirected graph" sense and not in the "pixels"
sense. Text-based format which seems pretty straightforward to parse (maybe
I should have a go at parsing it for fun).
Installed Bandage and viewed the results. Not sure what exactly I'm looking at,
but it works and looks pretty enough I guess. [This
example](https://github.com/rrwick/Bandage/wiki/Simple-example) was helpful to
get a sense of what it's trying to show me.
Unicycler has some configuration knobs to tweak:
> Unicycler can be run in three modes: conservative, normal (the default) and
> bold, set with the --mode option. Conservative mode is least likely to produce
> a complete assembly but has a very low risk of misassembly. Bold mode is most
> likely to produce a complete assembly but carries greater risk of misassembly.
> Normal mode is intermediate regarding both completeness and misassembly risk.
Reran with both conservative and bold modes and looked at the difference in the
results for the sample data. They're not the same, but I can't visually detect
any major obvious differences. Maybe it's not a big deal on this sample.
Once I got that running in a shell, I got it ported into Snakemake, shaving
a bunch of yaks along the way.
I noticed that Snakemake can take a JSON config file instead of YAML. Fantastic,
switched over to that right away:
{
"containers": {
"fastqc": "docker://staphb/fastqc:0.12.1",
"unicycler": "docker://staphb/unicycler:0.5.0"
},
"samples": {
"short_read_example": [
"sample_data/short_reads_1.fastq.gz",
"sample_data/short_reads_2.fastq.gz"
]
}
}
Got the containers downloading via snakemake as well, so it's snakes all the way
down:
rule containers:
input:
expand("containers/{name}.sif", name=config["containers"].keys()),
rule container:
output:
"containers/{name}.sif",
params:
source=lambda wc: config["containers"][wc.name],
shell:
"singularity pull {output} {params.source}"
Got `snakefmt` working with Neoformat so I can `F6` in Vim to reformat. Had to
fuck around with the config because it was just emptying out the file — I think
the key was that:
* `snakefmt` edits in-place by default (gross).
* Need to use `replace`: `1` in the Neoformat config to deal with this.
And finally we can assemble:
def get_input_fastqs(wildcards):
return config["samples"][wildcards.sample]
rule unicycler_assemble:
log:
"logs/unicycler_assemble/{sample}.log",
container:
"containers/unicycler.sif"
threads: 8
input:
"containers/unicycler.sif",
get_input_fastqs,
output:
"assemblies/{sample}/assembly.gfa",
"assemblies/{sample}/assembly.fasta",
shell:
logged(
"/unicycler/unicycler-runner.py"
" --threads {threads}"
" --short1 {input[0]}"
" --short2 {input[1]}"
" --out assemblies/{wildcards.sample}/"
)
Puttered around changing my StumpWM and terminal colors/borders/etc a bit to
make them a little easier on my eyes. We'll see if it sticks.
## 2023-09-06
HG545 in the morning. Mostly understood things.
Asked the professor after about a question I had while reading the paper. One
of the things the paper did to confirm the region of interest was to use a PAC
(P1-derived artificial chromosome) to "rescue" the golden embryos. The
resulting fish showed mosaic rescue, confirming that the wild-type gene was
likely on that PAC, i.e. in the region of interest.
What I didn't understand is how injecting the plasmid into the embryos resulted
in the expression of the genes farther down the developmental line, e.g. does it
somehow get incorporated into the cells' genomes? It turns out to be messy.
First: you don't inject "a PAC" into the embryos, you inject "a shitload of
copies of the PAC" into the embryo. So all of the embryo's cells will have
copies of the PAC floating around inside. As the cell divides, those will get
diluted in daughter cells over time. Some of these copies will, by chance, make
it into the nucleus of their cells. And some of those (rarely) will get
randomly incorporated into the cell's genome, and from then on mitosis takes
over and the gene gets propagated normally. So the mosaic region from that
point forward will have the gene (and if the region happens to contain some
melanophores, also the rescued wild-type phenotype).
Got my Armis access at some point during class, so it's time to figure out how
to log into the various HPC clusters today.
Doubled checked exam schedule to make sure nothing conflicts. I think it's
fine.
Changed my school password after the network clusterfuck last week. Sigh.
Wanted to print something in the lab, realized I never installed any printing
support on this laptop, lol. `apt install cups` will hopefully Just Work. CUPS
interface is at `http://localhost:631/`. It did not Just Work. Surely 2024
will be The Year of Linux on the Desktop. Printer wouldn't configure itself,
driver didn't appear in the list when I tried to manually configure it through
CUPS. `apt install printer-driver-all` got me more drivers but not this one.
Tracked it down on the brother site and downloaded some `.deb` packages but
they're 32-bit instead of 64. Gave up at this point, what a janky shitshow of
an OS. If only everything else didn't suck in even worse ways.
Tried getting the VPN running. Installed with the script into `/opt/cisco`
(good). Got mysterious errors when trying to connect. Tried the GUI connection
manager, which runs, but gave a more informative error message I could search
for. Looks like I need to install `libwebkit2gtk-4.0-37`. Installed that, now
I get the login screen, but I can't 2FA because I only have Yubikeys set up but
that requires a real browser, not this jank webkitgtk thing, so now I need to
set up *another* 2FA method just for this. Good god. Tried to add a new 2FA
device via Duo, but that requires 2FAing *again*, but *this* time it's through
the Duo site which doesn't actually fucking work on Linux, so I can't add it
here, I'll need to use the Windows shitbox at home. God, I hate two-factor
authentication so much. It's *always* miserable.
Tried to do the homework for BIOINF-500 (creating a pubmed search alert). To do
this you need an NCBI account. Tried to log in via my UM account and managed to
500 the NCBI site. Incredible. Poked around and eventually got it working (I
think)? Created the alert, took a screenshot. Uploaded the PNG into Canvas for
the assignment. Canvas shows an error trying to retrieve it. Uploaded it to
the Canvas "My Files" and used *that* to submit the assignment, instead of
uploading the file directly, and that worked. *Incredible*. Why does *nothing*
ever work correctly?
Came home, tried to add the 2FA with the Windows box, but it failed in the same
way (hanging on the popup after successfully touching the yubikey). But
I finally figured out a workaround (in retrospect, I vaguely recall having to do
this when I added the extra yubikeys originally): log out of everything, then go
to log back into something (e.g. Wolverine Access), but **before** you 2FA in
*that* login process there will be a link to add a new device on the left side
of the screen. That will then require you to 2FA, but doing it *here* does
actually work. It seems absolutely wild to me that you need to *not* be logged
in if you want to manage 2FA, but here we are.
Now that I have the Duo app thing on my phone and connected, logging into the
VPN seems to work great. Yak shaved successfully.
## 2023-09-07
BS521 and its lab this morning.
Got my dotfiles synced to GL. One tricky thing: my remote `.bash_profile`
sources `/etc/profile` if it exists, but that causes problems on the cluster
because there's some read-only variable set in there that it doesn't like.
Commented out that line and everything looks okay. `ControlMaster` does work,
so I won't have to auth a billion times a day (thank god).
BS521 lab. Of course the AC isn't working so it's a billion degrees, lovely.
Installing tidyverse failed with inscrutable errors. After some googling it
[looks like](https://blog.zenggyu.com/posts/en/2018-01-29-installing-r-r-packages-e-g-tidyverse-and-rstudio-on-ubuntu-linux/index.html)
there's *extra* dependencies you need to install on Linux (C programming is
wonderful):
sudo apt install libcurl4-openssl-dev libssl-dev libxml2-dev \
libfontconfig1-dev libharfbuzz-dev libfribidi-dev libfreetype6-dev \
libpng-dev libtiff5-dev libjpeg-dev
Still fucked, so I manually trawled through the tons of log output, googled
around more, and found that I need to configure an env variable:
Sys.setenv(PKG_CONFIG_PATH="/usr/lib/x86_64-linux-gnu/pkgconfig")
And finally it works. Started going through the lab stuff but then time was up
— will come back to it later.
Moving on to rotation lab work. Working from home today, so I wanted to get my
laptop working with my external monitor and keyboard and such. Had to swap out
the USB-C cable I was using previously because apparently that one doesn't work
for video ("universal" serial bus my ass) and then adjust my StumpWM `xrandr`
commands, but I did get it working, so now I can use the laptop at my desk,
which is nice.
Note to self: I can probably use the text-mode VPN UI now that I've got the 2FA
sorted out, and maybe remove the webkit crap I installed for it originally.
See lab notebook.
Finished BS521 lab 0. Wasn't too bad once I shaved the tidyverse yak. Getting
it written up with Latex required more Latex derusting, this time for code
listings and images (pulled some of this from the MS thesis). First, preamble
stuff:
% Listing package for code listings.
\RequirePackage{listings}
\lstdefinestyle{default}{
basicstyle=\footnotesize\ttfamily,
showtabs=true,
frame=lines,
aboveskip=10pt,
}
\lstset{
language=,
style=default,
}
% Used to embed plots.
\usepackage{graphicx}
% No paragraph indentation for homework, just looks awkward.
\setlength{\parindent}{0pt}
% Inline code.
\def\code#1{\small\texttt{#1}}
I really need to split these up into actual files I can include instead of
copypasting them a million times, but maybe I'll wait for one more practice
round first.
Usage:
% Code listings.
\begin{lstlisting}
prop.table(table(DATA$Race))
Black MexicanAmerican Other OtherHispanic White
0.22921790 0.23954747 0.04230202 0.03885883 0.45007378
\end{lstlisting}
% Graphics.
\includegraphics[]{figures/bmi-hist}
\begin{center}
\includegraphics[width=0.45\textwidth]{figures/hist-age}
\includegraphics[width=0.45\textwidth]{figures/hist-log-age}
\end{center}
To actually save PDF plots with R:
pdf("figures/foo.pdf", height=6, width=6)
hist(DATA$Foo, main = "Distribution of Foo", xlab="Foo")
dev.off()
Went to the poster session. Lots of stuff I don't understand, and a tiny bit
that I do.
## 2023-09-08
HG545 this morning.
Papers never say what *could* have gone wrong with what they did — you have to
just read between the lines and actively think about that (and what it would
have meant, and what you would have done if it did).
Learned about nonsense-mediated decay: a mechanism where mRNA with premature
stop codons is degraded, instead of expressing a (probably truncated) protein.
Without this, if you have a mutation that creates a stop codon in the middle of
the gene, you would see truncated protein expressed. But because of NMD, the
mRNA is degraded and doesn't express the broken protein (as much). This is good
not only to reduce wasted translation, but because the truncated proteins can be
actively bad.
One important control that was left out of the study where they wanted to find
where in the organism the target gene is being expressed: inject a probe with
GFP that intentionally shouldn't match *anything*, and expect it to show up
vaguely all over (or not at all).
Another example covered during class: if you suspected a phenotype was caused by
a mutation in a promoter (instead of in an exon), how would you test this?
There were a couple of things folks came up with:
* Could sequence the region in the mutant and wild-type population, compare to
see if the mutation segregates the two reliably.
* Old school: "reporter genes". I'm a bit fuzzy on this, but I think you insert
the promoter into a vector with some easily observable gene (e.g. luciferase,
a bioluminescent protein). Then you see if that product is expressed more or
less with the different variants of the promoter. This is a bit janky because
just yanking the promoter completely out of context can be problematic (e.g.
loses the chromatin structure around it, nearby enhancers/repressors, etc).
* Could use RNAseq to see if the mutants with the variants are producing more of
the RNA for that gene.
* Could use CHIPseq, if you know the transcription factors that bind to that
promoter. Fix, fragment, attach antibodies to the TFs, precipitate them out,
unfix, extract the DNA (all the remaining is whatever was bound to the
transcription factors), and then do the sequencing. You would expect to see
a larger signal if the mutation in the promoter is causing transcription
factors to be more likely to bind.
Got back and tried VPN'ing with the command-line client. It seemed to hang
after entering my password, but then I realized it had just silently tried to
2FA with my phone and I didn't notice. Trying again and being ready with Duo
let me log in, so I think I can probably ditch the webkit crap I installed for
the graphical thing.
See lab notebook.
Desktop machine wouldn't take input from my USB hub all of a sudden. Found some
bullshit in the logs, probably not worth debugging Yet More Linux Jank if I'm
just going to wipe this machine and install Debian on it soon anyway. Tried to
reboot and systemd hung at the end, so I just powercycled the damn thing. If
I could just have one single day where no computer broke for me, that would be
so nice.
Flu shots are available, need to get one so PI doesn't get pinged all the time.
Read for BS521 class. All still pretty basic. Cleaned up and turned in lab 0.
Finished homework 2 as well, just to get it out of the way. Or at least
I thought I did, except there are apparently Surprise Questions™ not in the book
to do with R. I'll do that this weekend.
## 2023-09-09
Actually finished BS521 homework 2. Realized my Latex `\code` shortcut was
broken:
% Broken, doesn't scope the \small so later text is changed.
\def\code#1{\small\texttt{#1}}
% Fixed v v
\def\code#1{{\small\texttt{#1}}}
Did a first draft of the HG545 assignment 1. This one is a lot harder than the
stats homework. Need to polish it up and submit it tomorrow.
## 2023-09-10
Polished and submitted the HG545 homework. We'll see how it goes, I guess.
## 2023-09-11
HG545 discussion. This paper was pretty straightforward.
Lab. See lab notebook.
Met with PIBS peer mentor.
HG545 second paper was posted, need to do an initial read of that tonight.
## 2023-09-12
BS521 again. Mostly basic linear regression stuff, but got a few interesting
tidbits out of it, mostly about the coefficient of determination, also called
`R²` or `r²`. This is the square of the correlation coefficient `r`, and it is
said to mean "the fraction of the variability in the data that is explained by
the linear model". So an `r²` of `0.7` would mean "70% of the variation in the
data is explained by the model".
*Intuitively* what this means would be to look at the total variability in the
data, i.e.:
(- (reduce #'max y) (reduce #'min y))
Then convert the data to residuals by subtracting out the model:
(mapcar #'- (mapcar #'model x) y)
and look at home much variability remains:
(- (reduce #'max residuals) (reduce #'min residuals))
Compare the two to see the fraction that remains after accounting for the model.
Looked into some "R for actual programmers" resources so maybe I can feel like
I'm flailing less:
* <https://arrgh.tim-smith.us/>
* <https://r4ds.hadley.nz/>
* <https://adv-r.hadley.nz/>
* <https://www.burns-stat.com/documents/books/the-r-inferno/>
* <https://www.burns-stat.com/documents/tutorials/impatient-r/>
* <https://www.burns-stat.com/documents/books/tao-te-programming/>
Lunch at a place called Maizie's. Was actually pretty good!
Doing Yet Another Round of Paperwork for the VA. So much red tape. Did what
I could here, but there's a bunch I can't do until I get home after class today.
See lab notebook.
So far I'm loving the look of the stumpwm config changes I made the other day.
Shouldn't have waited this long to clean things up. TODO: use
`select-from-menu` to implement a better screen-switching shortcut in stump.
Figured out how to print. Use <https://mprint.umich.edu/maps?sites> to find
a reasonable printer nearby, then Print Here to use it. You upload a PDF or
whatever through the web UI. Good enough, it works. One color paper cost
$3.22 of my (apparent) $24 print budget. Welp.
Spent a bit more time tracking down my white whale font from that 1979 Science
issue. Identifont came to the rescue and I think I finally have an answer, or
at least something very close: "Rotation" by Arthur Ritzel from 1971.
Unfortunately a 50-year old font still has ghastly licensing options, so I'll
probably never be able to *use* it, but at least I have peace of mind, I guess.