Google Bigtable
Bigtable is a key-value and wide-column store, ideal for fast access to structured, semi-structured, or unstructured data. Extend your database application to build AI-powered experiences leveraging Bigtable's Langchain integrations.
This notebook goes over how to use Bigtable to save, load and delete langchain documents with BigtableLoader
and BigtableSaver
.
Learn more about the package on GitHub.
Before You Beginβ
To run this notebook, you will need to do the following:
- Create a Google Cloud Project
- Enable the Bigtable API
- Create a Bigtable instance
- Create a Bigtable table
- Create Bigtable access credentials
After confirmed access to database in the runtime environment of this notebook, filling the following values and run the cell before running example scripts.
# @markdown Please specify an instance and a table for demo purpose.
INSTANCE_ID = "my_instance" # @param {type:"string"}
TABLE_ID = "my_table" # @param {type:"string"}
π¦π Library Installationβ
The integration lives in its own langchain-google-bigtable
package, so we need to install it.
%pip install -upgrade --quiet langchain-google-bigtable
Colab only: Uncomment the following cell to restart the kernel or use the button to restart the kernel. For Vertex AI Workbench you can restart the terminal using the button on top.
# # Automatically restart kernel after installs so that your environment can access the new packages
# import IPython
# app = IPython.Application.instance()
# app.kernel.do_shutdown(True)
β Set Your Google Cloud Projectβ
Set your Google Cloud project so that you can leverage Google Cloud resources within this notebook.
If you don't know your project ID, try the following:
- Run
gcloud config list
. - Run
gcloud projects list
. - See the support page: Locate the project ID.
# @markdown Please fill in the value below with your Google Cloud project ID and then run the cell.
PROJECT_ID = "my-project-id" # @param {type:"string"}
# Set the project id
!gcloud config set project {PROJECT_ID}
π Authenticationβ
Authenticate to Google Cloud as the IAM user logged into this notebook in order to access your Google Cloud Project.
- If you are using Colab to run this notebook, use the cell below and continue.
- If you are using Vertex AI Workbench, check out the setup instructions here.
from google.colab import auth
auth.authenticate_user()
Basic Usageβ
Using the saverβ
Save langchain documents with BigtableSaver.add_documents(<documents>)
. To initialize BigtableSaver
class you need to provide 2 things:
instance_id
- An instance of Bigtable.table_id
- The name of the table within the Bigtable to store langchain documents.
from langchain_core.documents import Document
from langchain_google_bigtable import BigtableSaver
test_docs = [
Document(
page_content="Apple Granny Smith 150 0.99 1",
metadata={"fruit_id": 1},
),
Document(
page_content="Banana Cavendish 200 0.59 0",
metadata={"fruit_id": 2},
),
Document(
page_content="Orange Navel 80 1.29 1",
metadata={"fruit_id": 3},
),
]
saver = BigtableSaver(
instance_id=INSTANCE_ID,
table_id=TABLE_ID,
)
saver.add_documents(test_docs)
API Reference:
Querying for Documents from Bigtableβ
For more details on connecting to a Bigtable table, please check the Python SDK documentation.
Load documents from tableβ
Load langchain documents with BigtableLoader.load()
or BigtableLoader.lazy_load()
. lazy_load
returns a generator that only queries database during the iteration. To initialize BigtableLoader
class you need to provide:
instance_id
- An instance of Bigtable.table_id
- The name of the table within the Bigtable to store langchain documents.
from langchain_google_bigtable import BigtableLoader
loader = BigtableLoader(
instance_id=INSTANCE_ID,
table_id=TABLE_ID,
)
for doc in loader.lazy_load():
print(doc)
break
Delete documentsβ
Delete a list of langchain documents from Bigtable table with BigtableSaver.delete(<documents>)
.
from langchain_google_bigtable import BigtableSaver
docs = loader.load()
print("Documents before delete: ", docs)
onedoc = test_docs[0]
saver.delete([onedoc])
print("Documents after delete: ", loader.load())
Advanced Usageβ
Limiting the returned rowsβ
There are two ways to limit the returned rows:
import google.cloud.bigtable.row_filters as row_filters
filter_loader = BigtableLoader(
INSTANCE_ID, TABLE_ID, filter=row_filters.ColumnQualifierRegexFilter(b"os_build")
)
from google.cloud.bigtable.row_set import RowSet
row_set = RowSet()
row_set.add_row_range_from_keys(
start_key="phone#4c410523#20190501", end_key="phone#4c410523#201906201"
)
row_set_loader = BigtableLoader(
INSTANCE_ID,
TABLE_ID,
row_set=row_set,
)
Custom clientβ
The client created by default is the default client, using only admin=True option. To use a non-default, a custom client can be passed to the constructor.
from google.cloud import bigtable
custom_client_loader = BigtableLoader(
INSTANCE_ID,
TABLE_ID,
client=bigtable.Client(...),
)
Custom contentβ
The BigtableLoader assumes there is a column family called langchain
, that has a column called content
, that contains values encoded in UTF-8. These defaults can be changed like so:
from langchain_google_bigtable import Encoding
custom_content_loader = BigtableLoader(
INSTANCE_ID,
TABLE_ID,
content_encoding=Encoding.ASCII,
content_column_family="my_content_family",
content_column_name="my_content_column_name",
)
Metadata mappingβ
By default, the metadata
map on the Document
object will contain a single key, rowkey
, with the value of the row's rowkey value. To add more items to that map, use metadata_mapping.
import json
from langchain_google_bigtable import MetadataMapping
metadata_mapping_loader = BigtableLoader(
INSTANCE_ID,
TABLE_ID,
metadata_mappings=[
MetadataMapping(
column_family="my_int_family",
column_name="my_int_column",
metadata_key="key_in_metadata_map",
encoding=Encoding.INT_BIG_ENDIAN,
),
MetadataMapping(
column_family="my_custom_family",
column_name="my_custom_column",
metadata_key="custom_key",
encoding=Encoding.CUSTOM,
custom_decoding_func=lambda input: json.loads(input.decode()),
custom_encoding_func=lambda input: str.encode(json.dumps(input)),
),
],
)
Metadata as JSONβ
If there is a column in Bigtable that contains a JSON string that you would like to have added to the output document metadata, it is possible to add the following parameters to BigtableLoader. Note, the default value for metadata_as_json_encoding
is UTF-8.
metadata_as_json_loader = BigtableLoader(
INSTANCE_ID,
TABLE_ID,
metadata_as_json_encoding=Encoding.ASCII,
metadata_as_json_family="my_metadata_as_json_family",
metadata_as_json_name="my_metadata_as_json_column_name",
)
Customize BigtableSaverβ
The BigtableSaver is also customizable similar to BigtableLoader.
saver = BigtableSaver(
INSTANCE_ID,
TABLE_ID,
client=bigtable.Client(...),
content_encoding=Encoding.ASCII,
content_column_family="my_content_family",
content_column_name="my_content_column_name",
metadata_mappings=[
MetadataMapping(
column_family="my_int_family",
column_name="my_int_column",
metadata_key="key_in_metadata_map",
encoding=Encoding.INT_BIG_ENDIAN,
),
MetadataMapping(
column_family="my_custom_family",
column_name="my_custom_column",
metadata_key="custom_key",
encoding=Encoding.CUSTOM,
custom_decoding_func=lambda input: json.loads(input.decode()),
custom_encoding_func=lambda input: str.encode(json.dumps(input)),
),
],
metadata_as_json_encoding=Encoding.ASCII,
metadata_as_json_family="my_metadata_as_json_family",
metadata_as_json_name="my_metadata_as_json_column_name",
)