There have been some discussions lately on gender ratios and equality in Computer Science. In this post I don’t want to rehash the scientific studies, but just talk about my own experience as a man in the field of theoretical computer science, ever since I started graduate school nearly 20 years ago. I don’t presume to talk for all males, but I don’t think my story is completely non-representative. This is also a good opportunity to recommend that people read the research life stories series that Omer Reingold initiated on this blog, as well as the Turing centennial series on Luca Trevisan’s blog.

Like many other new grad students, when I just started at the Weizmann Institute of Science, I was extremely anxious. While I loved (the little I knew about) theory, I was not sure if I’m smart enough to do original research in this area. The professors seemed like beings from another world, but one thing that helped was that at least I did share some background with them. Almost all of them were males (and in fact, this being Israel, almost all were, like me, Ashkenazi secular jews). So, it was not completely unthinkable to imagine that I would some day be like them.

Another thing that helped me a lot in my early time in Weizmann was my study group.
I quickly found three friends, all male, and we spent many hours together studying for exams, talking about research, but also playing video games.

As I progressed in the field, I’ve had a great many research interactions and collaborations. I’ve met people at all hours and various locations, including coffee shops, restaurants, and private homes. I’ve never had to worry about the potential for sexual harassment, nor that I would be judged by my looks or what I wear. I am the type of person that speaks their opinion loudly, perhaps sometimes a little too loudly. In my experience, women that act the same are judged by a different standard and often “rub people the wrong way” and get tagged as having problematic personalities. Also, I was never afraid to ask potentially stupid questions. I did not need to fear that asking such a question would confirm people’s bias about me that I don’t belong in this field. Asking potentially stupid questions is one of the most useful ways to learn and evolve as a scientist.

Last but not least, I would most definitely not be where I am if my wife did not quit her job in Israel to move with me across the Atlantic for my postdoc. Also, while I think of myself as an involved father, I have to admit that Ravit still does more of the childcare duties. (As I’m reading this out loud, my daughter is commenting that I am not such an involved father…) Again, I am not saying that all male scientists are like me, nor that there are no female scientists that greatly rely on the support of their partners, but I don’t think my situation is atypical.

On more general terms, one misconception I see about science in such discussions is that it is about “things” or “facts” and not people, and hence perhaps should be free of biases.
But in fact science is an extremely social enterprise. Since graduating, I have never written a solo paper. Much of my work is spent talking to people in front of a whiteboard or notepad, and I’ve learned a great deal from discussions with fellow scientists. This also explains why certain subfields of science can have outlying gender or ethnicities ratios. People just feel more comfortable working with others similar to them, and role models greatly influence the fields students choose to pursue. For example, there is nothing innately Hungarian about combinatorics. Similarly, there is nothing innately feminine about cryptography, but rather a few extremely strong female scientists have had a huge effect. The influence of people such as Shafi Goldwasser, Tal Rabin, and others has not been limited just to their (amazing) research results but also as role models and mentors to many female cryptographers (and many male cryptographers as well, myself included).

I don’t like the expression “leveling the playing field” because Science is not a game. This is not about competing but about collaborating with each other to uncover the mysteries of the universe. But us males (especially those that, like me, come from amply represented ethnicities), should be aware and grateful for the advantages we’ve had in our careers. We should not try to remove these advantages: access to role models, freedom from bias, ease of networking, are all good things. We should however strive for a world where everyone enjoys the same benefits. In such a world we’ll have more scientists that are more productive and happier, and that will only be better for science.

Fall is coming, and with it the annual holiday of FOCS. FOCS 2017 will be held at Berkeley from Oct 15-17, with a day of workshops/tutorials on Oct 14th. Registrations are now open at http://focs17.simons.berkeley.edu/registration.html

Early registration deadline for the conference is Sept 25th, 2017.

Speaking of workshops, there is one way to guarantee that FOCS has a workshop in an area you care about, and it is to organize such a workshop yourself!

Speaking from experience, organizing a workshop is non-zero but not unmanageable amount of work. It is a great way to connect with colleagues in your area, as well as expose some other people in the TCS community to it. The call for workshops is now up at http://focs17.simons.berkeley.edu/workshopAndTutorial.html

A proposal for a workshop can be quite minimal – the main thing you need is the names of the organizers and speakers. To submit a proposal, send an email to Aleskander Mądry and James Lee by September 3, 2017.

One of the mysteries of computation is that, as far as we can tell, the time complexity of many natural computational problems is either $poly(n)$ (with a small exponent, such as $1,2$ or $3$) or at least $2^{\Omega(n)}$ (with a not too small coefficient in the exponent). This is perhaps the main reason why we are excited when people discover an $n^{20}$ (or even $n^{polylog(n)}$) time algorithm for an important problem X: we view this as evidence that X is of the EASY type, and the exponent is likely to be improved over time (as indeed most often happens).
Similarly, this is why we care about hardness reductions, even when they hide huge constants or exponents: we view them as evidence that a problem is of the HARD type, and expect (as is also often the case) the efficiency of the reduction to be eventually improved.
This is also why our asymptotic theory of efficiency in informative in a finite world, and we don’t spend too much time worrying about the fact that $n^{10} > 2^{n^{0.1}}$ for every input length $n$ that could possibly fit in the world’s storage capacity.

Of course there are results such as the time hierarchy theorem and Ladner’s theorem that tell us that “intermediate complexity” do exist, but the hope is that this doesn’t happen for “natural” problems.
(In particular, we can artificially create problems of complexity $2^{\sqrt{n}}$ or $n^{\log n}$ by padding exponentially hard problems. This is one of the reasons why Impagliazzo Paturi and Zane argued that for NP problems, the parameter $n$ should be thought of the solution size and not the input size.)

One family of natural problems are constraint satisfaction problems (CSPs). For some finite alphabet $\Sigma$ and constant $k$, a CSP $CSP(\mathcal{P})$ is characterized by a set $\mathcal{P}$ of predicates from $\Sigma^k \rightarrow \{0,1\}$.
The input for $CSP(\mathcal{P})$ is a collection of functions $P_1,\ldots,P_m :\Sigma^n \rightarrow \{0,1\}$, each of which applies some predicate in $\mathcal{P}$ to a $k$-subset of its coordinates.
The basic task in CSPs is exact satisfiability, where the goal is to tell whether there is some assignment $x$ for the inputs that such that $P_1(x)=\cdots = P_m(x)=1$.
The celebrated CSP dichotomy conjecture (which perhaps should now be called a theorem) implies (in its modern, algebraic form) that for every $\mathcal{P}$, either there is a polynomial-time (in fact I believe that at most $O(n^3)$ time) algorithm for the exact satisfiability of $CSP(\mathcal{P})$ or there is a constant-factor blowup reduction from 3SAT to $CSP(\mathcal{P})$, implying that (under the exponential-time hypothesis or ETH) its complexity is $2^{\Omega(n)}$.

The approximation problem for $CSP(\mathcal{P})$ is parameterized by two numbers $0 < s < c < 1$ (known as the soundness and completeness paramters) and asks to distinguish, given an input $P_1,\ldots,P_m$ of $CSP(\mathcal{P})$, between the case that there is some assignment $x$ such that $\tfrac{1}{m}\sum_{i=1}^m P_i(x) \geq c$, and the case that $\tfrac{1}{m}\sum_{i=1}^m P_i(x) \leq s$ for every $x\in \Sigma^n$. There is one sense in which understanding the complexity of the approximation problem is easier, since in the context of approximation, the pesky difference between 3SAT and 3XOR (bane of many failed NP vs P proofs, and the main reason the dichotomy conjecture is so hard to prove) disappears. On the other hand, in this context a problem such as BIPARTITNESS (or $CSP(\neq)$), which was in the EASY column of exactly solving, become the NP-hard problem of Max-Cut.

One of the most fascinating consequences of the Unique Games Conjecture is that, if it is true, then there are in fact CSPs for which their approximation is neither EASY nor HARD. That is, the Unique Games Conjecture (plus the ETH) implies the following conjecture:

Intermediate Complexity Conjecture: For every $\epsilon>0$, there exist some $\mathcal{P}$, $0 and $\delta,\mu >0$ such that $c$ vs $s$ approximation of $CSP(\mathcal{P})$ can be done in time $2^{n^\epsilon}$ but (the potentially easier) $c+\mu$ vs $s-\mu$  approximation cannot be done in time $2^{n^\delta}$.

This is a consequence of the subexponential time algorithm for unique games, which in particular implies that for, say, $c=1/10$, there is some $\epsilon = \epsilon(s)$ that tends to zero with $s$ such that the $s$ vs $c$ problem for the unique games CSP can be solved in time $2^{n^\epsilon}$.
On the other hand, the Unique Games Conjecture implies that for every $0, no matter how close $s$ is to zero, or $c$ is to $1$, the $s$ vs $c$ problem for unique games is NP hard and hence (under the ETH) cannot be solved in time $2^{n^{o(1)}}$.

I find this conjecture very fascinating. A priori, we might expect a zero one law for CSP’s complexity, where, depending on the approximation quality, one could either solve the problem in $2^{n^{o(1)}}$ time or require at least $2^{n^{1-o(1)}}$ time. Indeed, we have good reasons to believe that predicates such as 3XOR and 3SAT satisfy such a zero/one law. But, if the intermediate complexity conjecture is true then for some predicates (such as unique games itself, and probably even max cut, which can hardly be called “unnatural”) we would get a non-trivial curve of the running time vs approximation quality, that looks something like this:

Figure 1: Conjectured time complexity of $s$ vs $c=0.99$ approximation of unique games as a function of $s$ assuming the unique games conjecture. The UGC characterizes the threshold $\underline{s}>0$ when the time complexity becomes $2^{n^\epsilon}$ for some $\epsilon>0$, while there is also a (not yet fully determined) threshold $\overline{s} when there is a linear blowup reduction from 3SAT and hence (under the ETH) the complexity is $2^{n^{1-o(1)}}$.

(One other setting where people were not sure if the complexity of some finitely specified object must be either exponential or polynomial is group theory, where one can define the complexity of a finitely generated group as the number of distinct elements that can be generated by a word of length $n$. There it turned out there are groups of “intermediate complexity”.)

What is also quite interesting is that we might be able to prove the intermediate complexity conjecture without resolving the UGC. In an exciting recent work, Dinur, Khot, Kindler, Minzer and Safra gave a combinatorial hypothesis that, if true, would imply the NP hardness of $s$ vs $c$ approximation for unique games where $c=0.499$, and $s$ can be arbitrarily close to $0$. Thus, if their hypothesis is true then, combined with the subexponential time algorithm mentioned above, it would imply the intermediate complexity conjecture. (If the construction is improved to obtain completeness $c$ that is close to $1$ then of course it would prove the unique games conjecture as well.)

Through discussions with Pravesh Kothari and David Steurer, we showed that the DKKMS combinatorial hypothesis is equivalent to a fairly simple to state statement. Informally, it can be stated as follows: let $G$ be the graph on $n\times n$ binary matrices, such that two matrices $A,B$ have an edge if $A-B$ has rank one over GF(2). (UGC/PCP experts might recognize this graph as the degree two short code graph.) This graph is not a small set expander in the sense that, if you think about it for few minutes, you see that there are sets $S$ matrices of measure $o(1)$ such that if we pick a random $A\in S$ and a random neighbor $B$ of $A$, then with constant probability $B$ will also be in $S$. The DKKMS hypothesis is equivalent to the conjecture that these sets that you think about after few minutes encompass all such examples.

Let me state this more formally: Lets say that a subset $U \subseteq GF(2)^{n\times n}$ is a basic set if $U = \{ A : Au = v \}$ or $U = \{ A : u^\top A = v^\top \}$ where $u,v \in GF(2)^n$ are column vectors. Note that each such set has measure $2^{-n}$ among all matrices and that a random neighbor of an element in $U$ will be in $U$ with probability $1/2$. For every constant $c$, you can clearly construct a set where a random neighbor stays inside with probability $2^{-c}$ by intersecting $c$ basic sets. You can also keep this probability constant even if you just subsampled a constant fraction of this intersection (and you can add a few random vertices as well). The DKKMS hypothesis is equivalent to the statement that this is all you can do. That is, it is equivalent to the following statement:

Inverse short code conjecture: For every $\epsilon>0$ there exists some $\delta>0$ and $c\in \mathbb{N}$ such that if $n$ is large enough and $G=G_n$ is the graph above with vertex set $GF(2)^{n \times n}$ and $A\sim B$ if $rank(A+B)=1$, then for every $S \subseteq GF(2)^{n \times n}$, if $S$ satisfies $\Pr_{A \in S, B \sim A}[B\in S] \geq \epsilon$ then there are $c$ basic sets $U_1,\ldots, U_c$ such that $U = \cap_{i\in [c]} U_i$ satisfies $|S \cap U | \geq \delta |U|$.

I call this an “inverse conjecture” since it is aimed at characterizing the sets $S$ that have unusually small expansion in the short-code graphs in terms of their correlation with basic sets.
I find it a natural question, though have to admit that, despite thinking about it some time (and also talking about it, not just with David and Pravesh but also Irit Dinur, Shachar Lovett and Kaave Hoesseini) I still don’t have a strong intuition whether it’s false or true.
Given that the short code graph is a Cayley graph over the Boolean cube, this seems like a question that could be amenable to tools from Fourier analysis and additive combinatorics.
Some partial progress has been made by the DKKMS folks in their new manuscript.
In fact, the same observations show that their original hypothesis is also equivalent not just to the “inverse short code conjecture” but also to Hypothesis 1.7 in this report, which is an analogous “inverse conjecture” for the Grassmanian graph where, for $k \gg \ell$, the vertices are dimension $\ell$ subspaces of $GF(2)^k$, and two subspaces $V,V'$ are neighbors if their intersection has dimension $\ell-1$.

Nevertheless, at the moment this question is still wide open, and I hope this writeup encourages more people to work on it. Other than resolving this question, there are some other interesting research directions, including:

1. The DKKMS reduction combined with the Grigoriev 3XOR instance yields a concrete candidate sum-of-squares integrality gap for, say, $0.499$ vs $0.001$ unique games. By construction, the SOS value of this instance is $\geq 0.499$, but is analyzing the soundness of this instance easier than the general case?

2. The underlying building block for DKKMS is a problem known as a smooth label cover. This is not precisely a CSP, but can we show that it has intermediate complexity? i.e., give a subexponential time algorithm for it? If the DKKMS hypothesis is correct then we know that such an algorithm should be an SDP hierarchy.

3. Suppose the combinatorial hypothesis is true, is there an inherent reason why its proof should not be a low degree SOS proof? After all, related isoperimetric results on the short code graph have been shown by low degree SOS proofs. Also, is the proof of the current best bounds coming from the DKKMS manuscript SOS-izable?

4. The quantiative bounds for the DKKMS reduction are terrible. I haven’t worked out all the details but it seems that the reduction from 3XOR to $s$ vs $0.499$ approximation of unique games would map an $n$ variable instance to an instance of at least $n^{2^{1/s}}$ (or maybe even $n^{2^{2^{1/s}}})$) size. Is this inherent?

5. Is there a general conjecture that would predict for $CSP(\mathcal{P})$ (say even if $\mathcal{P}$ is a single predicate) and $0< s < c < 1$ what should be the time complexity of $s \pm o(1)$ vs $c \pm o(1)$ approximation for it? We have some approaches of trying to predict when the complexity should be $2^{n^{1-o(1)}}$ by considering pairwise independent distributions. I feel like a conjecture trying to characterize what can be done in $2^{n^{f(\epsilon)}}$ time would have something to do with defining a suitable notion of “$(1+\epsilon)$-wise independent distributions” perhaps via concepts of mutual information. For example, perhaps we can argue that a smooth label cover is a $1+\epsilon$-wise independent CSP to a certain extent.

TL;DR: New notes on introduction to theoretical computer science are available at http://www.introtcs.org.

This fall I will be teaching CS 121 at Harvard: Introduction to Theoretical Computer Science.
This type of “intro theory” course is taught at many universities, sometimes under the name “introduction to the theory of computation” or “computability and automata”, typically using Sipser’s excellent book.
Sipser’s book is incredibly well-written and liked by students, but there have been many developments in theoretical computer science and CS at large since it was written, both in terms of new results, techniques, and applications, as well as new emphases and points of view on this area.

So what should an “intro TCS” course contain these days?
My sense is that this course is first and foremost about studying computation in its own right, as opposed to being a tool to solve other problems.
This is not just about computability and complexity, but also about looking at different ways to measure the “goodness” of algorithms, especially given the drastic change in the way algorithms interact with the world these days.
So, for example, students in such a course should get at least some awareness of questions such as privacy or incentive-compatibility, even if doing these topics justice requires dedicated courses.

Another point that should be “hammered in” more is of hardness as a resource.
This of course arises in derandomization and cryptography, where recently cryptocurrencies have been quite explicit about equating computational difficulty with “mining” precious metals.
The interaction of computation with the physical world need to also be covered, and these days an “intro TCS” course cannot skip talking about quantum computing, which is our current best approach for modelling physically feasible computation.

But more than anything, considering computation as an object in its own right forces students to change their way of thinking, and truly come to terms with questions such as how do we model computation, how code is also data, and all the self-referential “paradoxes” and complications this entails.
This has always been present in “intro TOC” courses, but students can sometimes lose the big perspective while working on the n-th proof of non-regularity using the pumping lemma, or the m-th NP hardness reduction.

I am of course not alone in thinking of the need for a “more modern” intro TCS course.
See for example, Scott Aaronson’s MIT course. Also CMU’s renowned (or is it infamous? 🙂 ) Great Ideas in TCS has been following such an approach for quite a while, and the folks at CMU have been very kind with advice and suggestions.

There are also books such as Aaronson’s Quantum Computing Since Democritus and Moore and Mertens’ The nature of computation that take a similar perspective, though neither is quite appropriate as an introductory undergraduate textbook.

Thus, I am working on writing a new set of lecture notes, which are available on http://www.introtcs.org.
These notes are very much a work in progress, and I welcome comments and suggestions.
In particular, the source for these notes is posted on a github repository and people are very much invited to use the issues and pull requests features there to post suggestions, comments, and point out (or even fix!) typos and bugs.

Some of the differences between these notes and the typical treatment include the following:

• Very little automata: We will not start with automata. I find regular and context free languages to be important topics, but believe that they should be introduced as an example of a restricted model after one learns about general models and undecidability. For example, once you’ve seen the impossiblity of the halting problem, you will understand why operating systems typically allow you to search for files using regular expressions and not by specifying an arbitrary function, as you would not want the shell to enter into an infinite loop.
Generally, automata play a very small role in this version of this course.
• Start with straightline programs: Rather than talk about Turing machines, the first computational model is an extremely simple programming language that has only a single operation: foo := bar NAND baz assigns to the variable foo the result of the NAND of the variables bar and baz. There are no loops, if statements, or subroutines, so this corresponds to straightline programs that are of course the same as Boolean circuits. Since students are already familiar with programming, I believe that starting with such an ultra-simple programming language makes things more concrete. Moreover, thanks to the efforts of my wonderful teaching fellow Juan Esteller, there is an implementation of the NAND, NAND++, and NAND<< programming languages that are taught in this course.
• De-emphasize Turing machines: Rather than Turing machines, our universal computational model is an enhancement of the NAND programming language to the NAND++ programming language that allows loops by simply stipulating that the program execution goes back to the first line if the loop variable has the value 1. We also introduce an index variable i so that we can use foo_i to access a potentially unbounded memory. Of course, this is equivalent to Turing machines, and we show this equivalence (as well as equivalence to the lambda calculus and other models) . We also give a further enhancement of the NAND++ language that is equivalent to RAM programs, and so allows us to measure running time in the same it is measured in algorithms courses.
• Algorithms: While a theory of computation course is traditionally about the impossibility results or “lower bounds” part of TCS, it cannot afford to ignore algorithms completely. You can’t understand the challenge of trying to prove hardness before you’ve seen non-trivial algorithms. So, we will mention results such as the Karatsuba algorithm for multiplication, and how very often there is a fine line separating problems such as minimum cut and maximum cut, 2SAT and 3SAT, determinant and permanent, etc..
• Move fast and break things: Because we skip automata and don’t delve into the intricacies of Turing machines (and because our NAND model is essentially designed to make the proof of the Cook-Levin Theorem easy), we will (hopefully) be done with NP completeness before the mid-point of this course. This will allow us to talk about some advanced topics, see below.
• Randomness and cryptography: Randomness has become essential to computer science theory and practice, both as a computational resource and as a way to model the world. We talk about probabilistic algorithms, give some examples, and talk about derandomization (based on strong average-case assumptions), which is a nice seque into cryptography. In a cryptography lecture in an intro TCS course, people often show the RSA algorithm. We might do so too, but only as a way to motivate quantum computing later on. I want to convey the notion of how fantastic is the whole idea of public key encryption in the first place, and mention some of the even more fantastical notions such as fully homomorphic encryption. We will also see a public-key cryptosystem based on the hardness of solving noisy linear equations, which, unlike RSA, is probably going to be more similar to the commonly used public key systems 5-10 years from now. (Since I teach a whole course on cryptography, I might make the treatment of crypto in the TCS course briefer.)
• Algorithms and society: The TCS angle can and has provided useful insights on how we design algorithms where the input is provided by humans and their output effects humans. Students should at least know that this field exists, and mention issues such as incentive-compatibility, differential privacy and issues of governance in cryptocurrencies (perhaps tying this into the consensus problem in distributed computing).
• Entropy, coding, and compression: An understanding of entropy has become essential for both theoretical and applied computer science, whether it is to bound the complexity of a password, the number of samples required to learn a concept, or the length of a representation, entropy is all around us, and we will cover some basic information theory as well.
• Quantum computing: As mentioned, a theory of computer science has to talk about our realistic physical models for computation, and some of the surprising outcomes of quantum computation. While a full description of Shor’s algorithm may require too much lecture time, I find that Simon’s algorithm conveys the heart of its power and is much simpler to describe.

It is definitely an ambitious plan that will not make for an easy course to take or to teach.
Some might say that it’s a “recipe for disaster” but the saving grace is that, with the help of Salil Vadhan, we have assembled a wonderful teaching staff that should be a great help for the students in keeping up.
I will report back on how realistic it was at the end of the term.
Even with our fast pace, there are many areas of theory that are not mentioned above, but students should be exposed to. I hope to cover some of these topics in this course, or at least write lecture notes about them so that curious advanced students could read about on their own.

Once again, much of this is not novel, and courses such as the ones mentioned above have taken similar approaches. However, I still hope these notes can serve a useful resource for other people who teach similar courses.

Acknowledgements: These lecture notes are constantly evolving, and I am getting input from several people, for which I am deeply grateful.
Thanks to Jarosław Błasiok, Chi-Ning Chou, Juan Esteller, Alex Lombardi, Thomas Orton, Juan Perdomo, and Ryan Williams for helpful feedback.
Thanks to Anil Ada, Venkat Guruswami, and Ryan O’Donnell for helpful tips from their experience in teaching CMU 15-251.
Thanks to Juan Esteller for his work on implementing the NAND* languages.
Thanks to David Steurer for writing the scripts that we used for notes on sum of squares and that I am now repurposing to produce these notes and for his endless patience with me as I attempt to learn concepts such as git and makefiles. David’s code itself is based on several other packages, including pandoc, LateX, and the Tufte LaTeX and CSS packages.

(Guest post by Sanjeev Arora on behalf of the TheoryFest 2017 organizing committee)

Having entered into the organization of TheoryFest 2017 with some trepidation, we organizers were very relieved to see feedback such as  “Best. STOC. Ever”. This post shares with you the feedback we got from attendees, and our plans for the next couple of years.

Important: Next year’s Theory Fest will be June 25-29, 2018 in downtown LA. See you there!

## Some Impressions from the event

Location, location, location. Smack in the middle of lovely Montreal, in the midst of an ongoing street festival/party, and within walking distance to the waterfront and good restaurants. I wouldn’t complain if the conference were held here every year!

No catered lunches. This greatly reduced registration fees, and (together with the weak Canadian dollar) allowed us to be more generous with food and drink during breaks (including an open bar on three nights!).  It also encouraged attendees to explore Montreal a bit more.

Every STOC paper also presented at poster session: a hit. Billed as the most controversial aspect of the TheoryFest, this ended up (based upon in-person and online feedback) as the most positive change. Respondents said they expected to hate poster sessions and ended up loving them. Having experienced the energy of poster sessions at non-theory conferences, I am not surprised. (We head the complaint that the poster room was a bit noisy, and will address it.)

Energy level. A day filled with different kinds of events seems to help keep up people’s energy levels. Attendance at the talks was high; I saw hardly any people loitering about in the public areas during talks. Possibly, it helped that  STOC talks were in three parallel sessions; 50% more choice of topics at any given moment.

Highest attendance among STOC/FOCS held abroad. 377 attendees, which is quite high for the past decade.

Junior-senior lunches. Via a google doc page, 3 junior researchers (grads/postdocs) got to sign up to go to lunch with a senior researcher.  These lunches got rave reviews, though not enough people participated due to the short notice. Next time we’ll be better organized.

## Statistics from online survey

117 attendees have responded thus far to our online survey. Their average satisfaction with TheoryFest was 4.5/5 .

Large majorities (>75%) found the amount of time devoted to various sub-components “about right.”

In terms of quality, the poster session was the biggest hit, followed by the plenary long talks. STOC paper talks had somewhat weaker ratings. Here’s a more detailed feedback.

## Our responses to the most important feedback.

Given the general support for the changes introduced this year, next year’s event will see no more drastic changes; merely tweaks to improve the overall experience.

Some feedback we received include:

“Stoc talks need to be longer.” “Short stoc talks worked well; could shorten even more.”

The alternatives we see are:

• Longer 22 min talks in 4 parallel sessions instead of 3.
• Shorter 12 min talks in 2 parallel sessions (with poster sessions being the place to hear more details).

Both options have significant minority support but appear to lack majority support, which was the reason behind our split-the-baby decision of 17 min talks in 3 parallel sessions. We welcome further discussion. (The PC Chair Valerie King commented to me that she’d expected to dislike 17 min talks but ended up preferring this format.)
“The middle three days were a bit tiring; stretching from 9am to 10:30pm.”

We realize this and don’t see a good alternative. We hew to the philosophy that a big-tent theory conference should offer a large buffet of content, and attendees are free to decide how much to partake. One major problem with the old setup of STOC was an artificial cap on content —and arguably the day was nevertheless long and monotonous. Nobody complained about monotony this time.

“Some short plenary talks weren’t geared to a STOC audience.”

We’ll address  this by assigning each outside speaker a “theory envoy” who will give them feedback on their slides and talk plan. But clearly,  the problem will never go away 100%.

Overall, we welcome your feedback! If you didn’t fill out your Theory Fest survey thus far, please do so asap before we close it out, and feel free to also post your feedback as comments to this post.

The ITCS 2018 Call For Papers is now available!

ITCS is a conference that stands apart from all others. For a decade now, it has been celebrating the vibrancy and unity of our field of Theoretical Computer Science. See this blog post for a detailed discussion of what makes ITCS so cool and the brief description of ITCS’17 at the end of this post.

ITCS seeks to promote research that carries a strong conceptual message  (e.g., introducing a new concept, model or understanding, opening a new line of inquiry within traditional or interdisciplinary areas, introducing new mathematical techniques and methodologies, or new applications of known techniques). ITCS welcomes both conceptual and technical contributions whose contents will advance and inspire the greater theory community.

This year, ITCS will be held at MIT in Cambridge, MA from January 11-14, 2018.

The submission deadline is September 8, 2017, with notification of decisions by October 30, 2017.

Authors should strive to make their papers accessible not only to experts in their subarea, but also to the theory community at large. The committee will place a premium on writing that conveys clearly and in the simplest possible way what the paper is accomplishing.

Ten-page versions of accepted papers will be published in an electronic proceedings of the conference. However, the alternative of publishing a one page abstract with a link to a full PDF will also be available (to accommodate subsequent publication in journals that would not consider results that have been published in preliminary form in a conference proceedings).

You can find all the details in the official Call For Papers.

On last year’s ITCS (by the PC Chair Christos Papadimitriou)

This past ITCS (2017) was by all accounts the most successful ever.  We had 170+ submissions and 61 papers, including 5 “invited papers”, and 90+ registrants, all new records.  There was a voluntary poster session for authors to get a chance to go through more detail, and the famous Graduating Bits event, where the younger ones get their 5 minutes to show off their accomplishment and personality.

The spirit of the conference was invigorating, heartwarming, and great fun.  I believe none of the twelve sessions had fewer than 70 attendees — no parallelism, of course — while the now famous last session was among the best attended and went one hour overtime due to the excitement of discussion (compare with the last large conference that you attended).

Tomorrow is the last day of TheoryFest. My sense is that it was very successful in the most important metric that it was a great event for the people that attended it.  However, we will know more about this once we send out a questionnaire to attendees next week. (Please respond when you get it!)

Many people helped in making TheoryFest happen (my own role was quite limited: chairing the short invited paper committee), but two people that I think are worth singling out are the general chairs Hamed Hatami and Pierre McKenzie. Doing local arrangements often means a lot of work dealing with hotels, food, registrations, etc., without much recognition, but they really did an amazing job.

I might attempt later some longer expositions of some of the talks, but in the meantime, see Suresh’s and Dick’s posts. Luca writes about the awards. The best student paper awardee, Pasin Manurangsi, gave a wonderful talk about his almost tight hardness of approximation result for the classical densest-k-subgraph problem.  He managed to explain clearly what the result is, what are the limits of what we can hope for, and convey the main ideas of the proof, all in a 20 minute talk. The proof itself is beautiful: the reduction (from 3SAT) is simple and natural, but the analysis is quite subtle, combining clever new ideas with classical results in combinatorics.

Tomorrow is the workshop day which I am also looking forward to. Tselil Schramm said that we should probably have called our SOS workshop a “wutorial” since, while we will cover some “hot off the presses” results,  we are planning to mostly make the talks an accessible introduction to the field. In particular, Pablo Parrilo’s talk will be a broad overview of the sum of squares algorithm’s history, applications, and connections to other areas. (See my previous post for description of all the talks.) That said, there are some other great workshops in parallel to ours. That is probably one of the biggest problems with TheoryFest – too much great content and too little time..