Pub Date : 2026-01-29DOI: 10.48550/arxiv.2601.22344
Marc Aurèle Gilles, Heather Wilber
The low-rank approximation properties of Randomly Pivoted LU (RPLU), a variant of Gaussian elimination where pivots are sampled proportional to the squared entries of the Schur complement, are analyzed. It is shown that the RPLU iterates converge geometrically in expectation for matrices with rapidly decaying singular values. RPLU outperforms existing low-rank approximation algorithms in two settings: first, when memory is limited, RPLU can be implemented with $mathcal{O}(k^2 + m + n)$ storage and $mathcal{O}( k(m + n)+ kmathcal{M}(mat{A}) + k^3)$ operations, where $mathcal{M}(mat{A})$ is the cost of a matvec with $mat{A}inmathbb{C}^{ntimes m}$ or its adjoint, for a rank-$k$ approximation. Second, when the matrix and its Schur complements share exploitable structure, such as for Cauchy-like matrices. The efficacy of RPLU is illustrated with several examples, including applications in rational approximation and solving large linear systems on GPUs.
{"title":"Low-Rank Approximation by Randomly Pivoted LU","authors":"Marc Aurèle Gilles, Heather Wilber","doi":"10.48550/arxiv.2601.22344","DOIUrl":"https://doi.org/10.48550/arxiv.2601.22344","url":null,"abstract":"The low-rank approximation properties of Randomly Pivoted LU (RPLU), a variant of Gaussian elimination where pivots are sampled proportional to the squared entries of the Schur complement, are analyzed. It is shown that the RPLU iterates converge geometrically in expectation for matrices with rapidly decaying singular values. RPLU outperforms existing low-rank approximation algorithms in two settings: first, when memory is limited, RPLU can be implemented with $mathcal{O}(k^2 + m + n)$ storage and $mathcal{O}( k(m + n)+ kmathcal{M}(mat{A}) + k^3)$ operations, where $mathcal{M}(mat{A})$ is the cost of a matvec with $mat{A}inmathbb{C}^{ntimes m}$ or its adjoint, for a rank-$k$ approximation. Second, when the matrix and its Schur complements share exploitable structure, such as for Cauchy-like matrices. The efficacy of RPLU is illustrated with several examples, including applications in rational approximation and solving large linear systems on GPUs.","PeriodicalId":496270,"journal":{"name":"arXiv (Cornell University)","volume":"1 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2026-01-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147382012","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-11-13DOI: 10.48550/arxiv.2311.07133
Johnson, Joseph, Liu, Ricky Ini
We define a birational map between labelings of a rectangular poset and its associated trapezoidal poset. This map tropicalizes to a bijection between the plane partitions of these posets of fixed height, giving a new bijective proof of a result by Proctor. We also show that this map is equivariant with respect to birational rowmotion, resolving a conjecture of Williams and implying that birational rowmotion on trapezoidal posets has finite order.
{"title":"Plane partitions and rowmotion on rectangular and trapezoidal posets","authors":"Johnson, Joseph, Liu, Ricky Ini","doi":"10.48550/arxiv.2311.07133","DOIUrl":"https://doi.org/10.48550/arxiv.2311.07133","url":null,"abstract":"We define a birational map between labelings of a rectangular poset and its associated trapezoidal poset. This map tropicalizes to a bijection between the plane partitions of these posets of fixed height, giving a new bijective proof of a result by Proctor. We also show that this map is equivariant with respect to birational rowmotion, resolving a conjecture of Williams and implying that birational rowmotion on trapezoidal posets has finite order.","PeriodicalId":496270,"journal":{"name":"arXiv (Cornell University)","volume":"119 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-11-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"136352295","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
To mitigate the potential misuse of large language models (LLMs), recent research has developed watermarking algorithms, which restrict the generation process to leave an invisible trace for watermark detection. Due to the two-stage nature of the task, most studies evaluate the generation and detection separately, thereby presenting a challenge in unbiased, thorough, and applicable evaluations. In this paper, we introduce WaterBench, the first comprehensive benchmark for LLM watermarks, in which we design three crucial factors: (1) For textbf{benchmarking procedure}, to ensure an apples-to-apples comparison, we first adjust each watermarking method's hyper-parameter to reach the same watermarking strength, then jointly evaluate their generation and detection performance. (2) For textbf{task selection}, we diversify the input and output length to form a five-category taxonomy, covering $9$ tasks. (3) For textbf{evaluation metric}, we adopt the GPT4-Judge for automatically evaluating the decline of instruction-following abilities after watermarking. We evaluate $4$ open-source watermarks on $2$ LLMs under $2$ watermarking strengths and observe the common struggles for current methods on maintaining the generation quality. The code and data are available at url{https://github.com/THU-KEG/WaterBench}.
{"title":"WaterBench: Towards Holistic Evaluation of Watermarks for Large Language\u0000 Models","authors":"Tu, Shangqing, Sun, Yuliang, Bai, Yushi, Yu, Jifan, Hou, Lei, Li, Juanzi","doi":"10.48550/arxiv.2311.07138","DOIUrl":"https://doi.org/10.48550/arxiv.2311.07138","url":null,"abstract":"To mitigate the potential misuse of large language models (LLMs), recent research has developed watermarking algorithms, which restrict the generation process to leave an invisible trace for watermark detection. Due to the two-stage nature of the task, most studies evaluate the generation and detection separately, thereby presenting a challenge in unbiased, thorough, and applicable evaluations. In this paper, we introduce WaterBench, the first comprehensive benchmark for LLM watermarks, in which we design three crucial factors: (1) For textbf{benchmarking procedure}, to ensure an apples-to-apples comparison, we first adjust each watermarking method's hyper-parameter to reach the same watermarking strength, then jointly evaluate their generation and detection performance. (2) For textbf{task selection}, we diversify the input and output length to form a five-category taxonomy, covering $9$ tasks. (3) For textbf{evaluation metric}, we adopt the GPT4-Judge for automatically evaluating the decline of instruction-following abilities after watermarking. We evaluate $4$ open-source watermarks on $2$ LLMs under $2$ watermarking strengths and observe the common struggles for current methods on maintaining the generation quality. The code and data are available at url{https://github.com/THU-KEG/WaterBench}.","PeriodicalId":496270,"journal":{"name":"arXiv (Cornell University)","volume":"118 51","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-11-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"136352299","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}