aboutsummaryrefslogtreecommitdiff
path: root/vendor/encode_unicode/benches
diff options
context:
space:
mode:
Diffstat (limited to 'vendor/encode_unicode/benches')
-rw-r--r--vendor/encode_unicode/benches/multiiterators.rs93
1 files changed, 0 insertions, 93 deletions
diff --git a/vendor/encode_unicode/benches/multiiterators.rs b/vendor/encode_unicode/benches/multiiterators.rs
deleted file mode 100644
index 22c3eed..0000000
--- a/vendor/encode_unicode/benches/multiiterators.rs
+++ /dev/null
@@ -1,93 +0,0 @@
-// uses /usr/share/dict/ for text to convert to Vec<Utf*Char> and iterate over
-#![cfg(all(unix, feature="std"))]
-#![feature(test)]
-extern crate test;
-use test::{Bencher, black_box};
-#[macro_use] extern crate lazy_static;
-extern crate encode_unicode;
-use encode_unicode::{CharExt, Utf8Char, Utf16Char, iter_bytes, iter_units};
-
-static ENGLISH: &str = include_str!("/usr/share/dict/american-english");
-// TODO find a big chinese file; `aptitude search '?provides(wordlist)'` didn't have one
-lazy_static!{
- static ref UTF8CHARS: Vec<Utf8Char> = ENGLISH.chars().map(|c| c.to_utf8() ).collect();
- static ref UTF16CHARS: Vec<Utf16Char> = ENGLISH.chars().map(|c| c.to_utf16() ).collect();
-}
-
-
-#[bench]
-fn utf16_split_all_single_mulititerator(b: &mut Bencher) {
- b.iter(|| {
- iter_units(black_box(&*UTF16CHARS)).for_each(|u| assert!(u != 0) );
- });
-}
-#[bench]
-fn utf16_split_all_single_flatmap(b: &mut Bencher) {
- b.iter(|| {
- black_box(&*UTF16CHARS).iter().flat_map(|&u16c| u16c ).for_each(|u| assert!(u != 0) );
- });
-}
-#[bench]
-fn utf16_split_all_single_cloned_flatten(b: &mut Bencher) {
- b.iter(|| {
- black_box(&*UTF16CHARS).iter().cloned().flatten().for_each(|u| assert!(u != 0) );
- });
-}
-
-
-#[bench]
-fn utf8_split_mostly_ascii_multiiterator(b: &mut Bencher) {
- b.iter(|| {
- iter_bytes(black_box(&*UTF8CHARS)).for_each(|b| assert!(b != 0) );
- });
-}
-#[bench]
-fn utf8_split_mostly_ascii_flatmap(b: &mut Bencher) {
- b.iter(|| {
- black_box(&*UTF8CHARS).iter().flat_map(|&u8c| u8c ).for_each(|b| assert!(b != 0) );
- });
-}
-#[bench]
-fn utf8_split_mostly_ascii_cloned_flatten(b: &mut Bencher) {
- b.iter(|| {
- black_box(&*UTF8CHARS).iter().cloned().flatten().for_each(|b| assert!(b != 0) );
- });
-}
-
-
-#[bench]
-fn utf8_extend_mostly_ascii_multiiterator(b: &mut Bencher) {
- b.iter(|| {
- let vec: Vec<u8> = iter_bytes(black_box(&*UTF8CHARS)).collect();
- assert_eq!(black_box(vec).len(), ENGLISH.len());
- });
-}
-#[bench]
-fn utf8_extend_mostly_ascii_custom(b: &mut Bencher) {
- b.iter(|| {
- let vec: Vec<u8> = black_box(&*UTF8CHARS).iter().collect();
- assert_eq!(black_box(vec).len(), ENGLISH.len());
- });
-}
-#[bench]
-fn utf8_extend_mostly_ascii_custom_str(b: &mut Bencher) {
- b.iter(|| {
- let vec: String = black_box(&*UTF8CHARS).iter().cloned().collect();
- assert_eq!(black_box(vec).len(), ENGLISH.len());
- });
-}
-
-#[bench]
-fn utf16_extend_all_single_multiiterator(b: &mut Bencher) {
- b.iter(|| {
- let vec: Vec<u16> = iter_units(black_box(&*UTF16CHARS)).collect();
- assert!(black_box(vec).len() < ENGLISH.len());
- });
-}
-#[bench]
-fn utf16_extend_all_single_custom(b: &mut Bencher) {
- b.iter(|| {
- let vec: Vec<u16> = black_box(&*UTF16CHARS).iter().collect();
- assert!(black_box(vec).len() < ENGLISH.len());
- });
-}