Mathematik
Refine
Year of publication
- 1997 (16) (remove)
Document Type
- diplomthesis (5)
- Doctoral Thesis (4)
- Report (3)
- Preprint (2)
- Article (1)
- Conference Proceeding (1)
Has Fulltext
- yes (16)
Is part of the Bibliography
- no (16)
Keywords
- Cauchy-Anfangswertproblem (1)
- Dimension 2 (1)
- Dirichletsche L-Reihe; Nullstelle (1)
- Funktionenkörper ; Arithmetische Gruppe ; Auflösbare Gruppe ; Endlichkeit (1)
- Gitter <Mathematik> ; Basis <Mathematik> ; Reduktion ; Algorithmus ; Laufzeit ; L-unendlich-Norm ; Rucksackproblem ; Kryptosystem (1)
- Iteration (1)
- Laplace-Differentialgleichung (1)
- Message authentication (1)
- Security (1)
- Uniform resource locators (1)
Institute
- Mathematik (16)
- Informatik (6)
In der vorliegenden Arbeit untersuchen wir die Verteilung der Nullstellen Dirichletscher L-Reihen auf oder in der Nähe der kritischen Geraden. Diese Funktionen und ihre Nullstellen stehen im Mittelpunkt des Interesses bei einer Vielzahl klassischer zahlentheoretischer Fragestellungen; beispielsweise besagt die Verallgemeinerte Riemannsche Vermutung, daß sämtliche Nullstellen dieser Funktionen auf der kritischen Geraden liegen. Unsere Ergebnisse gehen unter anderem über die besten bislang bekannten Abschätzungen - für den Anteil der Nullstellen der Dirichletschen L-Reihen, die auf der kritischen Geraden liegen, - für den Anteil einfacher beziehungsweise m-facher Nullstellen sowie - über Nullstellen in der Nähe der kritischen Geraden hinaus. Wir setzen hiermit Arbeiten von A. Selberg, N. Levinson, J. B. Conrey und anderen fort und verallgemeinern Ergebnisse, die für die Riemannsche #-Funktion gültig sind, auf alle Dirichletschen LReihen beziehungsweise verbessern bisherige Resultate. Nach einer ausführlicheren Darstellung der Hintergründe zeigen wir einen Satz über Mittelwerte "geglätteter" L-Reihen, d.h. mit einem geeigneten Dirichlet-Polynom multiplizierte L-Reihen. Solche Mittelwertsätze stellen ein wesentliches Hilfsmittel zur Untersuchung der Nullstellenverteilung dar. Die in unserem Hauptsatz gegebene asymptotische Darstellung dieses Mittelwertes können wir dann nutzen, um die genannten Ergebnisse herzuleiten.
A memory checker for a data structure provides a method to check that the output of the data structure operations is consistent with the input even if the data is stored on some insecure medium. In [8] we present a general solution for all data structures that are based on insert(i,v) and delete(j) commands. In particular this includes stacks, queues, deques (double-ended queues) and lists. Here, we describe more time and space efficient solutions for stacks, queues and deques. Each algorithm takes only a single function evaluation of a pseudorandomlike function like DES or a collision-free hash function like MD5 or SHA for each push/pop resp. enqueue/dequeue command making our methods applicable to smart cards.
We show lower bounds for the signature size of incremental schemes which are secure against substitution attacks and support single block replacement. We prove that for documents of n blocks such schemes produce signatures of \Omega(n^(1/(2+c))) bits for any constant c>0. For schemes accessing only a single block resp. a constant number of blocks for each replacement this bound can be raised to \Omega(n) resp. \Omega(sqrt(n)). Additionally, we show that our technique yields a new lower bound for memory checkers.
We introduce the relationship between incremental cryptography and memory checkers. We present an incremental message authentication scheme based on the XOR MACs which supports insertion, deletion and other single block operations. Our scheme takes only a constant number of pseudorandom function evaluations for each update step and produces smaller authentication codes than the tree scheme presented in [BGG95]. Furthermore, it is secure against message substitution attacks, where the adversary is allowed to tamper messages before update steps, making it applicable to virus protection. From this scheme we derive memory checkers for data structures based on lists. Conversely, we use a lower bound for memory checkers to show that so-called message substitution detecting schemes produce signatures or authentication codes with size proportional to the message length.
We address to the problem to factor a large composite number by lattice reduction algorithms. Schnorr has shown that under a reasonable number theoretic assumptions this problem can be reduced to a simultaneous diophantine approximation problem. The latter in turn can be solved by finding sufficiently many l_1--short vectors in a suitably defined lattice. Using lattice basis reduction algorithms Schnorr and Euchner applied Schnorrs reduction technique to 40--bit long integers. Their implementation needed several hours to compute a 5% fraction of the solution, i.e., 6 out of 125 congruences which are necessary to factorize the composite. In this report we describe a more efficient implementation using stronger lattice basis reduction techniques incorporating ideas of Schnorr, Hoerner and Ritter. For 60--bit long integers our algorithm yields a complete factorization in less than 3 hours.
In this short note on my talk I want to point out the mathematical difficulties that arise in the study of the relation of Wightman and Euclidean quantum field theory, i.e., the relation between the hierarchies of Wightman and Schwinger functions. The two extreme cases where the reconstructed Wightman functions are either tempered distributions - the well-known Osterwalder-Schrader reconstruction - or modified Fourier hyperfunctions are discussed in some detail. Finally, some perpectives towards a classification of Euclidean reconstruction theorems are outlined and preliminary steps in that direction are presented.