Joshua Brody, S. K. Jakobsen, Dominik Scheder, P. Winkler
We consider the following cryptographic secret leaking problem. A group of players communicate with the goal of learning (and perhaps revealing) a secret held initially by one of them. Their conversation is monitored by a computationally unlimited eavesdropper, who wants to learn the identity of the secret-holder. Despite the unavailability of key, some protection can be provided to the identity of the secret-holder. We call the study of such communication problems, either from the group's or the eavesdropper's point of view, cryptogenography. We introduce a basic cryptogenography problem and show that two players can force the eavesdropper to missguess the origin of a secret bit with probability 1/3; we complement this with a hardness result showing that they cannot do better than than 3/8. We prove that larger numbers of players can do better than 0.5644, but no group of any size can achieve 0.75.
{"title":"Cryptogenography","authors":"Joshua Brody, S. K. Jakobsen, Dominik Scheder, P. Winkler","doi":"10.1145/2554797.2554800","DOIUrl":"https://doi.org/10.1145/2554797.2554800","url":null,"abstract":"We consider the following cryptographic secret leaking problem. A group of players communicate with the goal of learning (and perhaps revealing) a secret held initially by one of them. Their conversation is monitored by a computationally unlimited eavesdropper, who wants to learn the identity of the secret-holder. Despite the unavailability of key, some protection can be provided to the identity of the secret-holder. We call the study of such communication problems, either from the group's or the eavesdropper's point of view, cryptogenography. We introduce a basic cryptogenography problem and show that two players can force the eavesdropper to missguess the origin of a secret bit with probability 1/3; we complement this with a hardness result showing that they cannot do better than than 3/8. We prove that larger numbers of players can do better than 0.5644, but no group of any size can achieve 0.75.","PeriodicalId":382856,"journal":{"name":"Proceedings of the 5th conference on Innovations in theoretical computer science","volume":"61 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-01-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115003892","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
We derive structural constraints on the automorphism groups of strongly regular (s.r.) graphs, giving a surprisingly strong answer to a decades-old problem, with tantalizing implications to testing isomorphism of s.r. graphs, and raising new combinatorial challenges. S.r. graphs, while not believed to be Graph Isomorphism (GI) complete, have long been recognized as hard cases for GI, and, in this author's view, present some of the core difficulties of the general GI problem. Progress on the complexity of testing their isomorphism has been intermittent (Babai 1980, Spielman 1996, BW & CST (STOC'13) and BCSTW (FOCS'13)), and the current best bound is exp(Õ(n1/5)) (n is the number of vertices). Our main result is that if X is a s.r. graph then, with straightforward exceptions, the degree of the largest alternating group involved in the automorphism group Aut(X) (as a quotient of a subgroup) is O((ln n)2ln ln n). (The exceptions admit trivial linear-time GI testing.) The design of isomorphism tests for various classes of structures is intimately connected with the study of the automorphism groups of those structures. We include a brief survey of these connections, starting with an 1869 paper by Jordan on trees. In particular, our result amplifies the potential of Luks's divide-and-conquer methods (1980) to be applicable to testing isomorphism of s.r. graphs in quasipolynomial time. The challenge remains to find a hierarchy of combinatorial substructures through which this potential can be realized. We expect that the generality of our result will help in this regard; the result applies not only to s.r. graphs but to all graphs with strong spectral expansion and with a relatively small number of common neighbors for every pair of vertices. We state a purely mathematical conjecture that could bring us closer to finding the right kind of hierarchy. We also outline the broader GI context, and state conjectures in terms of "primitive coherent configurations." These are generalizations of s.r. graphs, relevant to the general GI problem. Another consequence of the main result is the strongest argument to date against GI-completeness of s.r. graphs: we prove that no polynomial-time categorical reduction of GI to isomorphism of s.r. graphs is possible. All known reductions between isomorphism problems of various classes of structures fit into our notion of "categorical reduction." The proof of the main result is elementary; it is based on known results in spectral graph theory and on a 1987 lemma on permutations by Ákos Seress and the author.
{"title":"On the automorphism groups of strongly regular graphs I","authors":"L. Babai","doi":"10.1145/2554797.2554830","DOIUrl":"https://doi.org/10.1145/2554797.2554830","url":null,"abstract":"We derive structural constraints on the automorphism groups of strongly regular (s.r.) graphs, giving a surprisingly strong answer to a decades-old problem, with tantalizing implications to testing isomorphism of s.r. graphs, and raising new combinatorial challenges. S.r. graphs, while not believed to be Graph Isomorphism (GI) complete, have long been recognized as hard cases for GI, and, in this author's view, present some of the core difficulties of the general GI problem. Progress on the complexity of testing their isomorphism has been intermittent (Babai 1980, Spielman 1996, BW & CST (STOC'13) and BCSTW (FOCS'13)), and the current best bound is exp(Õ(n1/5)) (n is the number of vertices). Our main result is that if X is a s.r. graph then, with straightforward exceptions, the degree of the largest alternating group involved in the automorphism group Aut(X) (as a quotient of a subgroup) is O((ln n)2ln ln n). (The exceptions admit trivial linear-time GI testing.) The design of isomorphism tests for various classes of structures is intimately connected with the study of the automorphism groups of those structures. We include a brief survey of these connections, starting with an 1869 paper by Jordan on trees. In particular, our result amplifies the potential of Luks's divide-and-conquer methods (1980) to be applicable to testing isomorphism of s.r. graphs in quasipolynomial time. The challenge remains to find a hierarchy of combinatorial substructures through which this potential can be realized. We expect that the generality of our result will help in this regard; the result applies not only to s.r. graphs but to all graphs with strong spectral expansion and with a relatively small number of common neighbors for every pair of vertices. We state a purely mathematical conjecture that could bring us closer to finding the right kind of hierarchy. We also outline the broader GI context, and state conjectures in terms of \"primitive coherent configurations.\" These are generalizations of s.r. graphs, relevant to the general GI problem. Another consequence of the main result is the strongest argument to date against GI-completeness of s.r. graphs: we prove that no polynomial-time categorical reduction of GI to isomorphism of s.r. graphs is possible. All known reductions between isomorphism problems of various classes of structures fit into our notion of \"categorical reduction.\" The proof of the main result is elementary; it is based on known results in spectral graph theory and on a 1987 lemma on permutations by Ákos Seress and the author.","PeriodicalId":382856,"journal":{"name":"Proceedings of the 5th conference on Innovations in theoretical computer science","volume":"30 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-01-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114299442","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
For a property P and a sub-property P', we say that P is P'-partially testable with q queries} if there exists an algorithm that distinguishes, with high probability, inputs in P' from inputs ε-far from P, using q queries. Some natural properties require many queries to test, but can be partitioned into a small number of subsets for which they are partially testable with very few queries, sometimes even a number independent of the input size. For properties over {0,1}, the notion of being thus partitionable ties in closely with Merlin-Arthur proofs of Proximity (MAPs) as defined independently in [14] a partition into r partially-testable properties is the same as a Merlin-Arthur system where the proof consists of the identity of one of the r partially-testable properties, giving a 2-way translation to an O(log r) size proof. Our main result is that for some low complexity properties a partition as above cannot exist, and moreover that for each of our properties there does not exist even a single sub-property featuring both a large size and a query-efficient partial test, in particular improving the lower bound set in [14]. For this we use neither the traditional Yao-type arguments nor the more recent communication complexity method, but open up a new approach for proving lower bounds. First, we use entropy analysis, which allows us to apply our arguments directly to 2-sided tests, thus avoiding the cost of the conversion in [14] from 2-sided to 1-sided tests. Broadly speaking we use "distinguishing instances" of a supposed test to show that a uniformly random choice of a member of the sub-property has "low entropy areas", ultimately leading to it having a low total entropy and hence having a small base set. Additionally, to have our arguments apply to adaptive tests, we use a mechanism of "rearranging" the input bits (through a decision tree that adaptively reads the entire input) to expose the low entropy that would otherwise not be apparent. We also explore the possibility of a connection in the other direction, namely whether the existence of a good partition (or MAP) can lead to a relatively query-efficient standard property test. We provide some preliminary results concerning this question, including a simple lower bound on the possible trade-off. Our second major result is a positive trade-off result for the restricted framework of 1-sided proximity oblivious tests. This is achieved through the construction of a "universal tester" that works the same for all properties admitting the restricted test. Our tester is very related to the notion of sample-based testing (for a non-constant number of queries) as defined by Goldreich and Ron in [13]. In particular it partially resolves an open problem raised by [13].
{"title":"Partial tests, universal tests and decomposability","authors":"E. Fischer, Yonatan Goldhirsh, Oded Lachish","doi":"10.1145/2554797.2554841","DOIUrl":"https://doi.org/10.1145/2554797.2554841","url":null,"abstract":"For a property P and a sub-property P', we say that P is P'-partially testable with q queries} if there exists an algorithm that distinguishes, with high probability, inputs in P' from inputs ε-far from P, using q queries. Some natural properties require many queries to test, but can be partitioned into a small number of subsets for which they are partially testable with very few queries, sometimes even a number independent of the input size. For properties over {0,1}, the notion of being thus partitionable ties in closely with Merlin-Arthur proofs of Proximity (MAPs) as defined independently in [14] a partition into r partially-testable properties is the same as a Merlin-Arthur system where the proof consists of the identity of one of the r partially-testable properties, giving a 2-way translation to an O(log r) size proof. Our main result is that for some low complexity properties a partition as above cannot exist, and moreover that for each of our properties there does not exist even a single sub-property featuring both a large size and a query-efficient partial test, in particular improving the lower bound set in [14]. For this we use neither the traditional Yao-type arguments nor the more recent communication complexity method, but open up a new approach for proving lower bounds. First, we use entropy analysis, which allows us to apply our arguments directly to 2-sided tests, thus avoiding the cost of the conversion in [14] from 2-sided to 1-sided tests. Broadly speaking we use \"distinguishing instances\" of a supposed test to show that a uniformly random choice of a member of the sub-property has \"low entropy areas\", ultimately leading to it having a low total entropy and hence having a small base set. Additionally, to have our arguments apply to adaptive tests, we use a mechanism of \"rearranging\" the input bits (through a decision tree that adaptively reads the entire input) to expose the low entropy that would otherwise not be apparent. We also explore the possibility of a connection in the other direction, namely whether the existence of a good partition (or MAP) can lead to a relatively query-efficient standard property test. We provide some preliminary results concerning this question, including a simple lower bound on the possible trade-off. Our second major result is a positive trade-off result for the restricted framework of 1-sided proximity oblivious tests. This is achieved through the construction of a \"universal tester\" that works the same for all properties admitting the restricted test. Our tester is very related to the notion of sample-based testing (for a non-constant number of queries) as defined by Goldreich and Ron in [13]. In particular it partially resolves an open problem raised by [13].","PeriodicalId":382856,"journal":{"name":"Proceedings of the 5th conference on Innovations in theoretical computer science","volume":"5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-01-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126938521","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Session details: Session 4: 16:00--16:10","authors":"David Xiao","doi":"10.1145/3255056","DOIUrl":"https://doi.org/10.1145/3255056","url":null,"abstract":"","PeriodicalId":382856,"journal":{"name":"Proceedings of the 5th conference on Innovations in theoretical computer science","volume":"17 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-01-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125366707","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Session details: Session 1: 08:30--8:40","authors":"Kobbi Nissim","doi":"10.1145/3255053","DOIUrl":"https://doi.org/10.1145/3255053","url":null,"abstract":"","PeriodicalId":382856,"journal":{"name":"Proceedings of the 5th conference on Innovations in theoretical computer science","volume":"37 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-01-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124384363","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
We prove new positive and negative results concerning the existence of truthful and individually rational mechanisms for purchasing private data from individuals with unbounded and sensitive privacy preferences. We strengthen the impossibility results of Ghosh and Roth (EC 2011) by extending it to a much wider class of privacy valuations. In particular, these include privacy valuations that are based on (ε δ)-differentially private mechanisms for non-zero δ, ones where the privacy costs are measured in a per-database manner (rather than taking the worst case), and ones that do not depend on the payments made to players (which might not be observable to an adversary). To bypass this impossibility result, we study a natural special setting where individuals have monotonic privacy valuations, which captures common contexts where certain values for private data are expected to lead to higher valuations for privacy (e. g. having a particular disease). We give new mechanisms that are individually rational for all players with monotonic privacy valuations, truthful for all players whose privacy valuations are not too large, and accurate if there are not too many players with too-large privacy valuations. We also prove matching lower bounds showing that in some respects our mechanism cannot be improved significantly.
{"title":"Redrawing the boundaries on purchasing data from privacy-sensitive individuals","authors":"Kobbi Nissim, S. Vadhan, David Xiao","doi":"10.1145/2554797.2554835","DOIUrl":"https://doi.org/10.1145/2554797.2554835","url":null,"abstract":"We prove new positive and negative results concerning the existence of truthful and individually rational mechanisms for purchasing private data from individuals with unbounded and sensitive privacy preferences. We strengthen the impossibility results of Ghosh and Roth (EC 2011) by extending it to a much wider class of privacy valuations. In particular, these include privacy valuations that are based on (ε δ)-differentially private mechanisms for non-zero δ, ones where the privacy costs are measured in a per-database manner (rather than taking the worst case), and ones that do not depend on the payments made to players (which might not be observable to an adversary). To bypass this impossibility result, we study a natural special setting where individuals have monotonic privacy valuations, which captures common contexts where certain values for private data are expected to lead to higher valuations for privacy (e. g. having a particular disease). We give new mechanisms that are individually rational for all players with monotonic privacy valuations, truthful for all players whose privacy valuations are not too large, and accurate if there are not too many players with too-large privacy valuations. We also prove matching lower bounds showing that in some respects our mechanism cannot be improved significantly.","PeriodicalId":382856,"journal":{"name":"Proceedings of the 5th conference on Innovations in theoretical computer science","volume":"9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-01-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121749227","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Quantum cryptography is based on the discovery that the laws of quantum mechanics allow levels of security that are impossible to replicate in a classical world [2, 8, 12]. Can such levels of security be guaranteed even when the quantum devices on which the protocol relies are untrusted? This fundamental question in quantum cryptography dates back to the early nineties when the challenge of achieving device independent quantum key distribution, or DIQKD, was first formulated [9]. We answer this challenge affirmatively by exhibiting a robust protocol for DIQKD and rigorously proving its security. The protocol achieves a linear key rate while tolerating a constant noise rate in the devices. The security proof assumes only that the devices can be modeled by the laws of quantum mechanics and are spatially isolated from each other and any adversary's laboratory. In particular, we emphasize that the devices may have quantum memory. All previous proofs of security relied either on the use of many independent pairs of devices [6, 4, 7], or on the absence of noise [10, 1]. To prove security for a DIQKD protocol it is necessary to establish at least that the generated key is truly random even in the presence of a quantum adversary. This is already a challenge, one that was recently resolved [14]. DIQKD is substantially harder, since now the protocol must also guarantee that the key is completely secret from the quantum adversary's point of view, and the entire protocol is robust against noise; this in spite of the substantial amounts of classical information leaked to the adversary throughout the protocol, as part of the error estimation and information reconciliation procedures. Our proof of security builds upon a number of techniques, including randomness extractors that are secure against quantum storage [3] as well as ideas originating in the coding strategy used in the proof of the Holevo-Schumacher-Westmoreland theorem [5, 11] which we apply to bound correlations across multiple rounds in a way not unrelated to information-theoretic proofs of the parallel repetition property for multiplayer games. Our main result can be understood as a new bound on monogamy [13] of entanglement in the type of complex scenario that arises in a key distribution protocol. Precise statements of our results and detailed proofs can be found at arXiv:1210.1810.
{"title":"Robust device independent quantum key distribution","authors":"U. Vazirani, Thomas Vidick","doi":"10.1145/2554797.2554802","DOIUrl":"https://doi.org/10.1145/2554797.2554802","url":null,"abstract":"Quantum cryptography is based on the discovery that the laws of quantum mechanics allow levels of security that are impossible to replicate in a classical world [2, 8, 12]. Can such levels of security be guaranteed even when the quantum devices on which the protocol relies are untrusted? This fundamental question in quantum cryptography dates back to the early nineties when the challenge of achieving device independent quantum key distribution, or DIQKD, was first formulated [9]. We answer this challenge affirmatively by exhibiting a robust protocol for DIQKD and rigorously proving its security. The protocol achieves a linear key rate while tolerating a constant noise rate in the devices. The security proof assumes only that the devices can be modeled by the laws of quantum mechanics and are spatially isolated from each other and any adversary's laboratory. In particular, we emphasize that the devices may have quantum memory. All previous proofs of security relied either on the use of many independent pairs of devices [6, 4, 7], or on the absence of noise [10, 1]. To prove security for a DIQKD protocol it is necessary to establish at least that the generated key is truly random even in the presence of a quantum adversary. This is already a challenge, one that was recently resolved [14]. DIQKD is substantially harder, since now the protocol must also guarantee that the key is completely secret from the quantum adversary's point of view, and the entire protocol is robust against noise; this in spite of the substantial amounts of classical information leaked to the adversary throughout the protocol, as part of the error estimation and information reconciliation procedures. Our proof of security builds upon a number of techniques, including randomness extractors that are secure against quantum storage [3] as well as ideas originating in the coding strategy used in the proof of the Holevo-Schumacher-Westmoreland theorem [5, 11] which we apply to bound correlations across multiple rounds in a way not unrelated to information-theoretic proofs of the parallel repetition property for multiplayer games. Our main result can be understood as a new bound on monogamy [13] of entanglement in the type of complex scenario that arises in a key distribution protocol. Precise statements of our results and detailed proofs can be found at arXiv:1210.1810.","PeriodicalId":382856,"journal":{"name":"Proceedings of the 5th conference on Innovations in theoretical computer science","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-01-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129147041","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Session details: Session 6: 10:30--10:40","authors":"V. Vaikuntanathan","doi":"10.1145/3255058","DOIUrl":"https://doi.org/10.1145/3255058","url":null,"abstract":"","PeriodicalId":382856,"journal":{"name":"Proceedings of the 5th conference on Innovations in theoretical computer science","volume":"55 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-01-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115896518","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
A random linear code has good minimal distance with high probability. The conjectured intractability of decoding random linear codes has recently found many applications in cryptography. One disadvantage of random linear codes is that their encoding complexity grows quadratically with the message length. Motivated by this disadvantage, we present a randomized construction of linear error-correcting codes which can be encoded in linear time and yet enjoy several useful features of random linear codes. Our construction is based on a linear-time computable hash function due to Ishai, Kushilevitz, Ostrovsky and Sahai [25]. We demonstrate the usefulness of these new codes by presenting several applications in coding theory and cryptography. These include the first family of linear-time encodable codes meeting the Gilbert-Varshamov bound, the first nontrivial linear-time secret sharing schemes, and plausible candidates for symmetric encryption and identification schemes which can be conjectured to achieve better asymptotic efficiency/security tradeoffs than all current candidates.
{"title":"Linear-time encodable codes meeting the gilbert-varshamov bound and their cryptographic applications","authors":"E. Druk, Y. Ishai","doi":"10.1145/2554797.2554815","DOIUrl":"https://doi.org/10.1145/2554797.2554815","url":null,"abstract":"A random linear code has good minimal distance with high probability. The conjectured intractability of decoding random linear codes has recently found many applications in cryptography. One disadvantage of random linear codes is that their encoding complexity grows quadratically with the message length. Motivated by this disadvantage, we present a randomized construction of linear error-correcting codes which can be encoded in linear time and yet enjoy several useful features of random linear codes. Our construction is based on a linear-time computable hash function due to Ishai, Kushilevitz, Ostrovsky and Sahai [25]. We demonstrate the usefulness of these new codes by presenting several applications in coding theory and cryptography. These include the first family of linear-time encodable codes meeting the Gilbert-Varshamov bound, the first nontrivial linear-time secret sharing schemes, and plausible candidates for symmetric encryption and identification schemes which can be conjectured to achieve better asymptotic efficiency/security tradeoffs than all current candidates.","PeriodicalId":382856,"journal":{"name":"Proceedings of the 5th conference on Innovations in theoretical computer science","volume":"9 11 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-01-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127043977","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Session details: Session 10: 10:30--10:40","authors":"Deeparnab Chakrabarty","doi":"10.1145/3255062","DOIUrl":"https://doi.org/10.1145/3255062","url":null,"abstract":"","PeriodicalId":382856,"journal":{"name":"Proceedings of the 5th conference on Innovations in theoretical computer science","volume":"39 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-01-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114547813","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}