| The Android Open Source Project | edbf3b6 | 2009-03-03 19:31:44 -0800 | [diff] [blame] | 1 | /* libs/opengles/Tokenizer.cpp | 
|  | 2 | ** | 
|  | 3 | ** Copyright 2006, The Android Open Source Project | 
|  | 4 | ** | 
|  | 5 | ** Licensed under the Apache License, Version 2.0 (the "License"); | 
|  | 6 | ** you may not use this file except in compliance with the License. | 
|  | 7 | ** You may obtain a copy of the License at | 
|  | 8 | ** | 
|  | 9 | **     http://www.apache.org/licenses/LICENSE-2.0 | 
|  | 10 | ** | 
|  | 11 | ** Unless required by applicable law or agreed to in writing, software | 
|  | 12 | ** distributed under the License is distributed on an "AS IS" BASIS, | 
|  | 13 | ** WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | 
|  | 14 | ** See the License for the specific language governing permissions and | 
|  | 15 | ** limitations under the License. | 
|  | 16 | */ | 
|  | 17 |  | 
|  | 18 | #include <stdio.h> | 
|  | 19 |  | 
|  | 20 | #include "Tokenizer.h" | 
|  | 21 |  | 
|  | 22 | // ---------------------------------------------------------------------------- | 
|  | 23 |  | 
|  | 24 | namespace android { | 
|  | 25 |  | 
|  | 26 | ANDROID_BASIC_TYPES_TRAITS(Tokenizer::run_t) | 
|  | 27 |  | 
|  | 28 | Tokenizer::Tokenizer() | 
|  | 29 | { | 
|  | 30 | } | 
|  | 31 |  | 
|  | 32 | Tokenizer::Tokenizer(const Tokenizer& other) | 
|  | 33 | : mRanges(other.mRanges) | 
|  | 34 | { | 
|  | 35 | } | 
|  | 36 |  | 
|  | 37 | Tokenizer::~Tokenizer() | 
|  | 38 | { | 
|  | 39 | } | 
|  | 40 |  | 
|  | 41 | uint32_t Tokenizer::acquire() | 
|  | 42 | { | 
|  | 43 | if (!mRanges.size() || mRanges[0].first) { | 
|  | 44 | _insertTokenAt(0,0); | 
|  | 45 | return 0; | 
|  | 46 | } | 
|  | 47 |  | 
|  | 48 | // just extend the first run | 
|  | 49 | const run_t& run = mRanges[0]; | 
|  | 50 | uint32_t token = run.first + run.length; | 
|  | 51 | _insertTokenAt(token, 1); | 
|  | 52 | return token; | 
|  | 53 | } | 
|  | 54 |  | 
|  | 55 | bool Tokenizer::isAcquired(uint32_t token) const | 
|  | 56 | { | 
|  | 57 | return (_indexOrderOf(token) >= 0); | 
|  | 58 | } | 
|  | 59 |  | 
|  | 60 | status_t Tokenizer::reserve(uint32_t token) | 
|  | 61 | { | 
|  | 62 | size_t o; | 
|  | 63 | const ssize_t i = _indexOrderOf(token, &o); | 
|  | 64 | if (i >= 0) { | 
|  | 65 | return BAD_VALUE; // this token is already taken | 
|  | 66 | } | 
|  | 67 | ssize_t err = _insertTokenAt(token, o); | 
|  | 68 | return (err<0) ? err : status_t(NO_ERROR); | 
|  | 69 | } | 
|  | 70 |  | 
|  | 71 | status_t Tokenizer::release(uint32_t token) | 
|  | 72 | { | 
|  | 73 | const ssize_t i = _indexOrderOf(token); | 
|  | 74 | if (i >= 0) { | 
|  | 75 | const run_t& run = mRanges[i]; | 
|  | 76 | if ((token >= run.first) && (token < run.first+run.length)) { | 
|  | 77 | // token in this range, we need to split | 
|  | 78 | run_t& run = mRanges.editItemAt(i); | 
|  | 79 | if ((token == run.first) || (token == run.first+run.length-1)) { | 
|  | 80 | if (token == run.first) { | 
|  | 81 | run.first += 1; | 
|  | 82 | } | 
|  | 83 | run.length -= 1; | 
|  | 84 | if (run.length == 0) { | 
|  | 85 | // XXX: should we systematically remove a run that's empty? | 
|  | 86 | mRanges.removeItemsAt(i); | 
|  | 87 | } | 
|  | 88 | } else { | 
|  | 89 | // split the run | 
|  | 90 | run_t new_run; | 
|  | 91 | new_run.first = token+1; | 
|  | 92 | new_run.length = run.first+run.length - new_run.first; | 
|  | 93 | run.length = token - run.first; | 
|  | 94 | mRanges.insertAt(new_run, i+1); | 
|  | 95 | } | 
|  | 96 | return NO_ERROR; | 
|  | 97 | } | 
|  | 98 | } | 
|  | 99 | return NAME_NOT_FOUND; | 
|  | 100 | } | 
|  | 101 |  | 
|  | 102 | ssize_t Tokenizer::_indexOrderOf(uint32_t token, size_t* order) const | 
|  | 103 | { | 
|  | 104 | // binary search | 
|  | 105 | ssize_t err = NAME_NOT_FOUND; | 
|  | 106 | ssize_t l = 0; | 
|  | 107 | ssize_t h = mRanges.size()-1; | 
|  | 108 | ssize_t mid; | 
|  | 109 | const run_t* a = mRanges.array(); | 
|  | 110 | while (l <= h) { | 
|  | 111 | mid = l + (h - l)/2; | 
|  | 112 | const run_t* const curr = a + mid; | 
|  | 113 | int c = 0; | 
|  | 114 | if (token < curr->first)                        c = 1; | 
|  | 115 | else if (token >= curr->first+curr->length)     c = -1; | 
|  | 116 | if (c == 0) { | 
|  | 117 | err = l = mid; | 
|  | 118 | break; | 
|  | 119 | } else if (c < 0) { | 
|  | 120 | l = mid + 1; | 
|  | 121 | } else { | 
|  | 122 | h = mid - 1; | 
|  | 123 | } | 
|  | 124 | } | 
|  | 125 | if (order) *order = l; | 
|  | 126 | return err; | 
|  | 127 | } | 
|  | 128 |  | 
|  | 129 | ssize_t Tokenizer::_insertTokenAt(uint32_t token, size_t index) | 
|  | 130 | { | 
|  | 131 | const size_t c = mRanges.size(); | 
|  | 132 |  | 
|  | 133 | if (index >= 1) { | 
|  | 134 | // do we need to merge with the previous run? | 
|  | 135 | run_t& p = mRanges.editItemAt(index-1); | 
|  | 136 | if (p.first+p.length == token) { | 
|  | 137 | p.length += 1; | 
|  | 138 | if (index < c) { | 
|  | 139 | const run_t& n = mRanges[index]; | 
|  | 140 | if (token+1 == n.first) { | 
|  | 141 | p.length += n.length; | 
|  | 142 | mRanges.removeItemsAt(index); | 
|  | 143 | } | 
|  | 144 | } | 
|  | 145 | return index; | 
|  | 146 | } | 
|  | 147 | } | 
|  | 148 |  | 
|  | 149 | if (index < c) { | 
|  | 150 | // do we need to merge with the next run? | 
|  | 151 | run_t& n = mRanges.editItemAt(index); | 
|  | 152 | if (token+1 == n.first) { | 
|  | 153 | n.first -= 1; | 
|  | 154 | n.length += 1; | 
|  | 155 | return index; | 
|  | 156 | } | 
|  | 157 | } | 
|  | 158 |  | 
|  | 159 | return mRanges.insertAt(run_t(token,1), index); | 
|  | 160 | } | 
|  | 161 |  | 
|  | 162 | void Tokenizer::dump() const | 
|  | 163 | { | 
|  | 164 | const run_t* ranges = mRanges.array(); | 
|  | 165 | const size_t c = mRanges.size(); | 
|  | 166 | LOGD("Tokenizer (%p, size = %u)\n", this, c); | 
|  | 167 | for (size_t i=0 ; i<c ; i++) { | 
|  | 168 | LOGD("%u: (%u, %u)\n", i, ranges[i].first, ranges[i].length); | 
|  | 169 | } | 
|  | 170 | } | 
|  | 171 |  | 
|  | 172 | }; // namespace android | 
|  | 173 |  |